2018
DOI: 10.1007/978-3-030-01234-2_1
|View full text |Cite
|
Sign up to set email alerts
|

CBAM: Convolutional Block Attention Module

Abstract: We propose Convolutional Block Attention Module (CBAM), a simple yet effective attention module for feed-forward convolutional neural networks. Given an intermediate feature map, our module sequentially infers attention maps along two separate dimensions, channel and spatial, then the attention maps are multiplied to the input feature map for adaptive feature refinement. Because CBAM is a lightweight and general module, it can be integrated into any CNN architectures seamlessly with negligible overheads and is… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

22
7,033
0
12

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 13,087 publications
(8,906 citation statements)
references
References 38 publications
22
7,033
0
12
Order By: Relevance
“…Zhang et al embedded edgeattention representations to guide the process of segmentation on optic disc, retinal vessel, and lung [40]. Attention modules incorporated in deep learning architectures have also shown their strengths in many computer vision based tasks [41], [42]. Schlemper et al encapsulated attention gates into a 3D U-Net architecture for abdominal organ segmentation [43].…”
Section: Previous Workmentioning
confidence: 99%
“…Zhang et al embedded edgeattention representations to guide the process of segmentation on optic disc, retinal vessel, and lung [40]. Attention modules incorporated in deep learning architectures have also shown their strengths in many computer vision based tasks [41], [42]. Schlemper et al encapsulated attention gates into a 3D U-Net architecture for abdominal organ segmentation [43].…”
Section: Previous Workmentioning
confidence: 99%
“…4, TIGN consists of three types of blocks, namely, attention, PN, and output blocks. Motivated by recent successful works on attention [43], [44], we introduce two attention blocks to focus on important appearance and motion features, individually. For each attention block, we adopt CBAM [44] and change the spatial attention to temporal attention.…”
Section: A Temporal Interval Generationmentioning
confidence: 99%
“…Indirect Supervision. We can also train the Selective Module in an unsupervised fashion [13,14]. Predicting detection results on each spatial location of specified feature maps is a key characteristic of one-stage detectors [1,12], and different gradients flow at different locations during training.…”
Section: Direct Supervision or Indirect Supervisionmentioning
confidence: 99%
“…Though DCN [12] uses the deformable convolution with spatial domain offsets to focus on the specific objects instead of the adjacent background, dispensable activations on the background locations will be also calculated. And many works [13,14,15] only use attention mechanisms to enhance certain features, which violates the original intention to decrease the size of search spaces.…”
Section: Introductionmentioning
confidence: 99%