2022
DOI: 10.1109/tgrs.2022.3194505
|View full text |Cite
|
Sign up to set email alerts
|

EMTCAL: Efficient Multiscale Transformer and Cross-Level Attention Learning for Remote Sensing Scene Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 52 publications
(32 citation statements)
references
References 62 publications
0
13
0
Order By: Relevance
“…On the other hand, the attention mechanism quantifies how much weight should be put into it. To date, the attention mechanism has been widely used in remote sensing, including scene classification [48], [49], object detection [50], [51], pansharpening [52], [53], and change detection [54], [55], and so forth. The essence of the attention-based deep neural network is to incorporate the similarity between different channels or positions to enhance pixel representations.…”
Section: A Affinity Modellingmentioning
confidence: 99%
“…On the other hand, the attention mechanism quantifies how much weight should be put into it. To date, the attention mechanism has been widely used in remote sensing, including scene classification [48], [49], object detection [50], [51], pansharpening [52], [53], and change detection [54], [55], and so forth. The essence of the attention-based deep neural network is to incorporate the similarity between different channels or positions to enhance pixel representations.…”
Section: A Affinity Modellingmentioning
confidence: 99%
“…Various transformers-based studies have been discussed in the literature 29 33 For the classification of remote sensing scenes, 29 a unique model of efficient multiscale transformer and cross-level attention learning was proposed.…”
Section: Related Workmentioning
confidence: 99%
“…Various transformers-based studies have been discussed in the literature 29 33 For the classification of remote sensing scenes, 29 a unique model of efficient multiscale transformer and cross-level attention learning was proposed. To obtain global visual features and rich contextual information from multiple features, this model used a multilayer feature extraction and contextual information extraction module, respectively.…”
Section: Related Workmentioning
confidence: 99%
“…Bazi et al [60] applied an attention mechanism to focus on different areas of the image and integrate global information. Tang et al [61] proposed a transformer that used multi-level features to mine the potential context information of remote sensing scenes. However, we believe that in complex landscapes, the transformer model has limitations for feature modeling and high computational complexity.…”
Section: Introductionmentioning
confidence: 99%