2021 IEEE Winter Conference on Applications of Computer Vision (WACV) 2021
DOI: 10.1109/wacv48630.2021.00333
|View full text |Cite
|
Sign up to set email alerts
|

Data-free Knowledge Distillation for Object Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 47 publications
(19 citation statements)
references
References 17 publications
0
14
0
Order By: Relevance
“…In the literature, Lopes et al proposes the first data-free approach for knowledge distillation, which utilizes statistical information of original training data to reconstruct a synthetic set during knowledge distillation (Lopes, Fenu, and Starner 2017). This seminal work has spawned several works, which has achieved impressive progressive on several tasks including detection (Chawla et al 2021), segmentation (Fang et al 2019), text classification (Ma et al 2020), graph classification (Deng and Zhang 2021) and Federated Learning (Zhu, Hong, and Zhou 2021). Despite the impressive progress, a vexing problem remains in DFKD, i.e., the inefficiency of data synthesis, which makes data-free training extraordinarily time-consuming.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In the literature, Lopes et al proposes the first data-free approach for knowledge distillation, which utilizes statistical information of original training data to reconstruct a synthetic set during knowledge distillation (Lopes, Fenu, and Starner 2017). This seminal work has spawned several works, which has achieved impressive progressive on several tasks including detection (Chawla et al 2021), segmentation (Fang et al 2019), text classification (Ma et al 2020), graph classification (Deng and Zhang 2021) and Federated Learning (Zhu, Hong, and Zhou 2021). Despite the impressive progress, a vexing problem remains in DFKD, i.e., the inefficiency of data synthesis, which makes data-free training extraordinarily time-consuming.…”
Section: Related Workmentioning
confidence: 99%
“…To learn a comparable student model, the synthetic set should contain sufficient samples to enable a comprehensive knowledge transfer from teachers. Consequentially, this poses a significant challenge to DFKD, since synthesizing a large-scale dataset is inevitably time-consuming, especially for sophisticated tasks like ImageNet recognition (Yin et al 2019) and COCO detection (Chawla et al 2021).…”
Section: Introductionmentioning
confidence: 99%
“…GAN-based methods [8,34,55,63] synthesized training samples through maximizing response on the discriminator. Prior-based methods [5] provide another perspective for data-free KD, where the synthetic data are forced to satisfy a pre-defined prior, such as total variance prior [3,36] and batch normalization statistics [5,8]. However, they all has the problem of mode collapse [6,45], so we propose a boundary-preserving intra-divergence loss for DeepInversion [56] to generate diverse samples.…”
Section: Related Workmentioning
confidence: 99%
“…[38,39] opt to use a large general proxy dataset to query the teacher, and their teacher outputs to on this data to train the student. Other methods [40,41,42,43] generate proxy data directly from the trained models, and use this data to train the students. Further, [44,45], also encourage generating samples the student and teacher disagree on.…”
Section: Related Workmentioning
confidence: 99%