2022
DOI: 10.3390/s22186939
|View full text |Cite
|
Sign up to set email alerts
|

Ghostformer: A GhostNet-Based Two-Stage Transformer for Small Object Detection

Abstract: In this paper, we propose a novel two-stage transformer with GhostNet, which improves the performance of the small object detection task. Specifically, based on the original Deformable Transformers for End-to-End Object Detection (deformable DETR), we chose GhostNet as the backbone to extract features, since it is better suited for an efficient feature extraction. Furthermore, at the target detection stage, we selected the 300 best bounding box results as regional proposals, which were subsequently set as prim… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(7 citation statements)
references
References 24 publications
0
2
0
Order By: Relevance
“…Finally, the eigenfeature map and Ghost feature map are combined to obtain the same number of feature maps at a lower computational cost [29]. Compared with current mainstream convolution operations, the Ghost Module has the following advantages [30,31]:…”
Section: Ghostnet Modelmentioning
confidence: 99%
“…Finally, the eigenfeature map and Ghost feature map are combined to obtain the same number of feature maps at a lower computational cost [29]. Compared with current mainstream convolution operations, the Ghost Module has the following advantages [30,31]:…”
Section: Ghostnet Modelmentioning
confidence: 99%
“…Another crucial element that impacts feature quality is the backbone network and its ability to extract both semantic and high-resolution features. GhostNet introduced in [93], offers a streamlined and more efficient network that delivers high-quality, multi-scale features to the transformer. Their Ghost module in this network partially generates the output feature map, with the remainder being recovered using simple linear operations.…”
Section: Improved Feature Representationmentioning
confidence: 99%
“…But YOLOv4 still has the problem of poor matching of manually designed anchor boxes on different tasks. Due to the excessive number of candidate boxes generated during the prediction process of one-stage methods [15][16][17][18][19][20], non-maximum suppression (NMS) processes are required to filter out a large number of candidate boxes, which not only reduces the inference speed but also fails to achieve truly end-to-end prediction, as shown in [21,22], due to the incorporation of a supervisory mechanism.…”
Section: Introductionmentioning
confidence: 99%