2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.00082
|View full text |Cite
|
Sign up to set email alerts
|

FcaNet: Frequency Channel Attention Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
192
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 604 publications
(270 citation statements)
references
References 25 publications
1
192
0
Order By: Relevance
“…Only using global average pooling in the squeeze module limits representational ability. To obtain a more powerful representation ability, Qin et al [57] rethought global information captured from the viewpoint of compression and analysed global average pooling in the frequency domain. They proved that global average pooling is a special case of the discrete cosine transform (DCT) and used this observation to propose a novel multi-spectral channel attention.…”
Section: Fcanetmentioning
confidence: 99%
See 1 more Smart Citation
“…Only using global average pooling in the squeeze module limits representational ability. To obtain a more powerful representation ability, Qin et al [57] rethought global information captured from the viewpoint of compression and analysed global average pooling in the frequency domain. They proved that global average pooling is a special case of the discrete cosine transform (DCT) and used this observation to propose a novel multi-spectral channel attention.…”
Section: Fcanetmentioning
confidence: 99%
“…Channel attention Generate attention mask across the channel domain and use it to select important channels. [5,25,37,[53][54][55][56][57][58][59][60] Spatial attention Generate attention mask across spatial domains and use it to select important spatial regions (e.g., [15,61]) or predict the most relevant spatial position directly (e.g., [7,31]). [8, 9, 15, 20-22, 26, 27, 31, 32, 34, 35, 41-47, 61-109] Temporal attention Generate attention mask in time and use it to select key frames.…”
Section: Introductionmentioning
confidence: 99%
“…Dai et al [34] proposed a second-order attention network to explore the feature correlations of intermediate layers for image super-resolution. Qin et al [35] proposed a novel multispectral channel attention involving the pre-processing of a channel attention mechanism in the frequency domain.…”
Section: Attention Mechanismsmentioning
confidence: 99%
“…Further, AANets [3] and SCNet [22] demonstrate how self-attention and self-calibration operations can augment standard convolutions, while GCNet [6] extends non-local neural networks to augment SE operations. More recently, prominent modules include ECA [35] which proposes one-dimensional convolutions to efficiently capture inter-channel interactions for channel attention and FCA [27] which proposes utilization of discrete cosine transform based frequency compression methods to effectively perform feature aggregation in SE in place of GAP.…”
Section: Related Work 21 Attention Modules For Cnnsmentioning
confidence: 99%
“…As a prominent method, Squeeze & Excitation (SE) [16] introduces channel attention modelling of global-average-pooled (GAP) feature representations, which is then enhanced by CBAM [37] through additional incorporation of spatial attention and utilization of both global-max-pooled (GMP) and GAP representations. Further, recent works [12,27,35] identify how channel attention can be made more efficient and effective, while a different direction of work augments convolutional operations with self-attention and calibration methods [3,22] to learn more effective feature representations.…”
Section: Introductionmentioning
confidence: 99%