2022
DOI: 10.1007/s00521-022-06953-8
|View full text |Cite
|
Sign up to set email alerts
|

Pooling in convolutional neural networks for medical image analysis: a survey and an empirical study

Abstract: Convolutional neural networks (CNN) are widely used in computer vision and medical image analysis as the state-of-the-art technique. In CNN, pooling layers are included mainly for downsampling the feature maps by aggregating features from local regions. Pooling can help CNN to learn invariant features and reduce computational complexity. Although the max and the average pooling are the widely used ones, various other pooling techniques are also proposed for differe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
50
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 87 publications
(50 citation statements)
references
References 124 publications
0
50
0
Order By: Relevance
“…The most used variants are: max or min pooling (higher or lower value goes to the input unmodified) and average pooling (an average value of all the input values affected is obtained as output). Nirthika et al [ 49 ] perform an empirical study of pooling operations in CNN for medical image analysis. They conclude that choosing an appropriate pooling technique for a particular job is related to the size and scale of the images and its class-specific features.…”
Section: Deep Learningmentioning
confidence: 99%
“…The most used variants are: max or min pooling (higher or lower value goes to the input unmodified) and average pooling (an average value of all the input values affected is obtained as output). Nirthika et al [ 49 ] perform an empirical study of pooling operations in CNN for medical image analysis. They conclude that choosing an appropriate pooling technique for a particular job is related to the size and scale of the images and its class-specific features.…”
Section: Deep Learningmentioning
confidence: 99%
“…In this paper, the selected output shape became 7 2 × 192, as the filter size f of 192 provides better adequacy than 96, but still less than 512, as the proposed method considers the importance of cost. By providing the specific kernel size k , strides s , pool size p, and other settings after the cut-points of each model, they eventually had a similar output shape of 7 2 × 192, making them compatible for a layer-wise fusion [21] . On the other hand, the auxiliary layers also contain an alpha DO with a rate of 0.2 that improves the regularization of the incoming fused features but does not affect the reshaping of each model's output.…”
Section: Layer-wise Fusion With Auxiliary Layersmentioning
confidence: 99%
“…On the other hand, the auxiliary layers also contain an alpha DO with a rate of 0.2 that improves the regularization of the incoming fused features but does not affect the reshaping of each model's output. If cases that other researchers or users use cut-points that do not reflect the ones in this proposed method, the Conv and AP settings will necessitate alterations to make them fit [21] .…”
Section: Layer-wise Fusion With Auxiliary Layersmentioning
confidence: 99%
“…7 is a very simple convolutional neural network (CNN) [23] model taking a spectrum of 18,000 values as input. The convolutional layer was made of 3 lters and a kernel size of 6, followed by a max pooling layer with pool size = 100 [24]. The atten output of the CNN was sent to two fully connected layers of respectively 512 and 1024 units.…”
Section: Machine Learning: Convolutional Neural Network Modelingmentioning
confidence: 99%