2022
DOI: 10.1016/j.patcog.2022.108899
|View full text |Cite
|
Sign up to set email alerts
|

Dimension-aware attention for efficient mobile networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 11 publications
0
1
0
Order By: Relevance
“…This helps to produce three feature matrices based on the dimension. Dimension-wise convolutions are used to encode spatial and channel-wise information in the input tensor efficiently [46,47]. The input tensor M has dimensions w, h, and d, where w represents width, d represents depth, and h represents height.…”
Section: ) Swin Transformer Trackmentioning
confidence: 99%
“…This helps to produce three feature matrices based on the dimension. Dimension-wise convolutions are used to encode spatial and channel-wise information in the input tensor efficiently [46,47]. The input tensor M has dimensions w, h, and d, where w represents width, d represents depth, and h represents height.…”
Section: ) Swin Transformer Trackmentioning
confidence: 99%
“…Existing attention mechanisms, such as CBAM and SE, typically employ global maximum pooling or global average pooling operations, which can result in the loss of spatial information. In contrast, the Coordinate Attention (CA) module [34] incorporates location information into channel attention, allowing for the consideration of both channel and location information. This integration effectively increases the emphasis on the target to be recognized within the image.…”
Section: Coordinate Attentionmentioning
confidence: 99%