2022
DOI: 10.7717/peerj-cs.1161
|View full text |Cite
|
Sign up to set email alerts
|

Enhanced mechanisms of pooling and channel attention for deep learning feature maps

Abstract: The pooling function is vital for deep neural networks (DNNs). The operation is to generalize the representation of feature maps and progressively cut down the spatial size of feature maps to optimize the computing consumption of the network. Furthermore, the function is also the basis for the computer vision attention mechanism. However, as a matter of fact, pooling is a down-sampling operation, which makes the feature-map representation approximately to small translations with the summary statistic of adjace… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 28 publications
(31 reference statements)
0
1
0
Order By: Relevance
“…As shown in Figure 8 b, a pooling layer is added between two multi-channel attention layers. Pooling layers integrate the features in a small adjacent area, preventing useless parameters from increasing the time complexity on the one hand, and enhancing the integration of features on the other [ 38 ].…”
Section: The Gait Prediction Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…As shown in Figure 8 b, a pooling layer is added between two multi-channel attention layers. Pooling layers integrate the features in a small adjacent area, preventing useless parameters from increasing the time complexity on the one hand, and enhancing the integration of features on the other [ 38 ].…”
Section: The Gait Prediction Modelmentioning
confidence: 99%
“…As shown in Figure 8b, a pooling layer is added between two multi-channel attention layers. Pooling layers integrate the features in a small adjacent area, preventing useless parameters from increasing the time complexity on the one hand, and enhancing the integration of features on the other [38]. The specific parameter settings for the decoder module are as follows: (1) the linear layers map the decomposed data dimension to (32,256,256), which is the same as the output of the encoder.…”
Section: Deep Multi-channel Attention Structurementioning
confidence: 99%