2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) 2019
DOI: 10.1109/iccvw.2019.00242
|View full text |Cite
|
Sign up to set email alerts
|

Understanding the Effects of Pre-Training for Object Detectors via Eigenspectrum

Abstract: ImageNet pre-training has been regarded as essential for training accurate object detectors for a long time. Recently, it has been shown that object detectors trained from randomly initialized weights can be on par with those finetuned from ImageNet pre-trained models. However, the effects of pre-training and the differences caused by pretraining are still not fully understood. In this paper, we analyze the eigenspectrum dynamics of the covariance matrix of each feature map in object detectors. Based on our an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
5
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 10 publications
(7 citation statements)
references
References 69 publications
2
5
0
Order By: Relevance
“…For example, in the right part of Figure 5, 500 fine-tuning iterations from DAP already achieve ≥ 65 AP .5 , while the corresponding CLS numbers are lower than 20. This demonstrates that a better pre-trained model can provide faster convergence speed, which is consistent with [25,20,16,34].…”
Section: Discussionsupporting
confidence: 73%
See 1 more Smart Citation
“…For example, in the right part of Figure 5, 500 fine-tuning iterations from DAP already achieve ≥ 65 AP .5 , while the corresponding CLS numbers are lower than 20. This demonstrates that a better pre-trained model can provide faster convergence speed, which is consistent with [25,20,16,34].…”
Section: Discussionsupporting
confidence: 73%
“…Classification pre-training may sometimes even harm localization when the downstream data is abundant while benefit classification [25]. Shinya et al try to understand the impact of ImageNet classification pre-training on detection and discover that the pre-trained model generates narrower eigenspectrum than the fromscratch model [34]. Recent work proposes a cheaper Montage pre-training for detection on the target detection data and obtains an on-par or better performance than ImageNet classification pre-training [48].…”
Section: Related Workmentioning
confidence: 99%
“…In particular, the number of singular values larger than a given (plausible) threshold changes consistently. A similar observation was also made in [53] in the context of object detectors (but based on the normalized eigenspectrum of the Hessian). This observation may explain the different dynamics of the optimization scheme for the pretrained model and the model trained from scratch.…”
Section: Spectral Evaluation Of Pretrainingsupporting
confidence: 67%
“…Epochs for the first learning rate decay, the second decay, and ending training are (8,11,12) for the 1× schedule, (16,22,24) for the 2× schedule, and (16,19,20) for the 20e schedule. To avoid overfitting by small learning rates [52], the 20e schedule is reasonable.…”
Section: D1 Common Settingsmentioning
confidence: 99%