2024
DOI: 10.1109/tpami.2023.3335410
|View full text |Cite
|
Sign up to set email alerts
|

DN-DETR: Accelerate DETR Training by Introducing Query DeNoising

Feng Li,
Hao Zhang,
Shilong Liu
et al.

Abstract: We present in this paper a novel denoising training method to speed up DETR (DEtection TRansformer) training and offer a deepened understanding of the slow convergence issue of DETR-like methods. We show that the slow convergence results from the instability of bipartite graph matching which causes inconsistent optimization goals in early training stages. To address this issue, except for the Hungarian loss, our method additionally feeds GT bounding boxes with noises into the Transformer decoder and trains the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2025
2025

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(1 citation statement)
references
References 26 publications
0
1
0
Order By: Relevance
“…Conditional DETR [19] introduces a conditioning mechanism on the decoder's attention, allowing for more flexible and targeted learning, which helps the model to better understand complex scenes and detect objects more accurately. DN-DETR [20] improves upon the deformable DETR by integrating a denoising strategy that enhances the model's robustness and accuracy, particularly in cluttered environments.…”
Section: Transformer-based Object Detection Technologymentioning
confidence: 99%
“…Conditional DETR [19] introduces a conditioning mechanism on the decoder's attention, allowing for more flexible and targeted learning, which helps the model to better understand complex scenes and detect objects more accurately. DN-DETR [20] improves upon the deformable DETR by integrating a denoising strategy that enhances the model's robustness and accuracy, particularly in cluttered environments.…”
Section: Transformer-based Object Detection Technologymentioning
confidence: 99%