2023
DOI: 10.1016/j.eswa.2023.119997
|View full text |Cite
|
Sign up to set email alerts
|

ETAM: Ensemble transformer with attention modules for detection of small objects

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(3 citation statements)
references
References 39 publications
0
3
0
Order By: Relevance
“…In addition to IoU, it also considers aspect ratio and centre distance. ETAM (Zhang, Xia, et al, 2023) strives to extract the subtle features of small objects without sacrificing the ability to detect larger objects. A Magnifying Glass (MG) module and Quadruple Attention (QA) are used to focus on the representative features of small objects and to enrich the features of small objects with width and height outside the channel and position, respectively.…”
Section: Related Workmentioning
confidence: 99%
“…In addition to IoU, it also considers aspect ratio and centre distance. ETAM (Zhang, Xia, et al, 2023) strives to extract the subtle features of small objects without sacrificing the ability to detect larger objects. A Magnifying Glass (MG) module and Quadruple Attention (QA) are used to focus on the representative features of small objects and to enrich the features of small objects with width and height outside the channel and position, respectively.…”
Section: Related Workmentioning
confidence: 99%
“…Zhang et al. [35] present the ensemble transformer with attention modules (ETAM) encoder, a powerful approach for detecting small objects by extracting subtle features. This method leads to significant improvements in small object detection performance across multiple datasets.…”
Section: Related Workmentioning
confidence: 99%
“…Thirdly, the maxpooling operation in the SPP module can cause significant information loss, which makes it difficult to obtain local and global information for localization and also results in a significant loss of computation. Fourthly, the information paths connecting different components in the YOLOv5 architecture limit its computational efficiency and are not optimal for feature extraction of small-scale objects [21,22].…”
Section: Yolov5 Basic Network Structurementioning
confidence: 99%