2021
DOI: 10.48550/arxiv.2103.09136
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

QueryDet: Cascaded Sparse Query for Accelerating High-Resolution Small Object Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 59 publications
0
3
0
Order By: Relevance
“…Nevertheless, with the threshold set to 0.5:0.95, the mAP value of the improved YOLOv5 algorithm achieves 34.0%, surpassing all other models. DMNet [38] 47.6% 28.2% YOLOv3 [39] 36.6% 17.5% Cascade-RCNN [40] 45.9% 24.3% YOLOv5l 46.4% 28.1% QueryDet [41] 48.2% 28.3% ClusDet [42] 53.2% 30.4% PRNet [43] 53.9% 32.0% Model of this paper 52.5% 34.0%…”
Section: Comparative Experimentsmentioning
confidence: 96%
“…Nevertheless, with the threshold set to 0.5:0.95, the mAP value of the improved YOLOv5 algorithm achieves 34.0%, surpassing all other models. DMNet [38] 47.6% 28.2% YOLOv3 [39] 36.6% 17.5% Cascade-RCNN [40] 45.9% 24.3% YOLOv5l 46.4% 28.1% QueryDet [41] 48.2% 28.3% ClusDet [42] 53.2% 30.4% PRNet [43] 53.9% 32.0% Model of this paper 52.5% 34.0%…”
Section: Comparative Experimentsmentioning
confidence: 96%
“…In the above subsection, the effectiveness of the proposed FFCLC has been verified, now, we choose some representative algorithms to compare with the FFCLC in different datasets of SAR-Ship-Dataset and SSDD. The detection results of Faster R-CNN, SSD, YOLOv3, TPH-YOLOv5 [49], QueryDet [50], YOLOv8, YOLOv5, and FFCLC in four different scenarios of SAR-Ship-Dataset are given in Fig. 12, and Fig.…”
Section: E Comparative Experimentsmentioning
confidence: 99%
“…In the commonly adopted implementation of feature pyramid-based dense detectors, the predictions of different layers are "flatten and concatenated" for loss computation, making the supervision on different layers to be conjectural. However, as pointed out in [49], the sample distribution is significantly imbalanced between different layers, because the feature size grows up in quadratic as the feature resolution increases. As a result, the training samples will be dominated by samples in lower levels, making the high-level samples lack of supervision, which will harm the performance on large objects.…”
Section: Supervision Conjunction In Fpnmentioning
confidence: 99%
“…VGG-16 0.967 0.959 0.912 AInnoFace [53] ResNet-152 0.970 0.961 0.918 RetinaFace [8] ResNet-152 0.969 0.961 0.918 RefineFace [54] ResNet-152 0.972 0.962 0.920 DSFD [18] ResNet-152 0.966 0.957 0.904 ASFD-D6 [52] ResNet-152 0.972 0.965 0.925 HAMBox [25] ResNet-50 0.970 0.964 0.933 TinaFace [62] ResNet 7. Firstly, we test the Linear-Reweight proposed in [49], in which the factors that linearly increase from 1.0 to 2.0 are assigned to each FPN level. Secondly, considering that most negative samples are greatly suppressed by the focal loss, we propose a vanilla re-weight strategy where the weight of each FPN layer is proportionate to the number of positive samples, which is denoted as Sum-Reweight.…”
Section: What Is the Best Label Assignment Disentanglement?mentioning
confidence: 99%