2022
DOI: 10.1109/lgrs.2022.3151353
|View full text |Cite
|
Sign up to set email alerts
|

High Resolution SAR Image Classification Using Global-Local Network Structure Based on Vision Transformer and CNN

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 25 publications
(15 citation statements)
references
References 12 publications
0
8
0
Order By: Relevance
“…Therefore, we introduce a classification layer after the hash network layer to improve the feature extraction capability and at the same time enhance the class discrimination capability of the model. In single-label image classification scenarios [45][46], we often use the classical cross-entropy loss (CE), which is denoted as:…”
Section: B Hash Learningmentioning
confidence: 99%
“…Therefore, we introduce a classification layer after the hash network layer to improve the feature extraction capability and at the same time enhance the class discrimination capability of the model. In single-label image classification scenarios [45][46], we often use the classical cross-entropy loss (CE), which is denoted as:…”
Section: B Hash Learningmentioning
confidence: 99%
“…But these models have higher complexity, which limits their applicability. To overcome this issue, work in [7] proposes design of global-local network structure with CNN, which assists in improving classification accuracy for multiple use cases. It uses vision transformers, which makes it useful for high density feature extraction & classification deployments.…”
Section: Literature Reviewmentioning
confidence: 99%
“…2. Some techniques that are effective in identifying synthetic images require significant computational resources, which limits their scalability for real-world applications [7,8,10]. 3.…”
Section: Issues With Existing Techniquesmentioning
confidence: 99%
“…ViT 31 has successfully applied the transformer 30 from natural language processing to computer vision 32 34 It demonstrates that transformer also has a strong ability to model spatial correlations of images. The main reason for the success of the transformer is its core component, the self-attention module that captures global information.…”
Section: Related Workmentioning
confidence: 99%