2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00874
|View full text |Cite
|
Sign up to set email alerts
|

Dreaming to Distill: Data-Free Knowledge Transfer via DeepInversion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
130
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 295 publications
(160 citation statements)
references
References 29 publications
2
130
0
Order By: Relevance
“…Some researchers have looked at optimizing an image map to maximize the activations of an individual or set of neurons [140]. Yin et al [141] generate images from activations for the sake of providing a knowledge distillation [142] set for compressing the network. Riberio et al [143] propose the LIME framework to sample around a data point to provide analysis of what caused the prediction.…”
Section: Interpretabilitymentioning
confidence: 99%
“…Some researchers have looked at optimizing an image map to maximize the activations of an individual or set of neurons [140]. Yin et al [141] generate images from activations for the sake of providing a knowledge distillation [142] set for compressing the network. Riberio et al [143] propose the LIME framework to sample around a data point to provide analysis of what caused the prediction.…”
Section: Interpretabilitymentioning
confidence: 99%
“…1 A classification problem over 10 classes is aimed for each dataset. We compare our RDSKD method with DAFL [9], MSKD where the denominator of L DS is replaced by the mode seeking loss in [10], data-free KD via DeepInversion (DeepI) [18], and zero-shot KD (ZSKD) [19]. For DeepI and ZSKD, image samples are directly generated from the teacher without a separate generator.…”
Section: Models and Parameter Settingsmentioning
confidence: 99%
“…It is also more challenging than training the generator and student simultaneously. Most works in the literature also introduced additional hyperparameters [9,17,18], requiring more efforts for parameter tuning. Leveraging and extending these preceding works, we develop RDSKD that is free from original data or meta-data while not requiring to fine-tune hyperparameters, which is in stark contrast to DAFL [9].…”
Section: Introductionmentioning
confidence: 99%
“…Specifically for super-resolution, generative adversarial network has been proven to be an advanced algorithm that can extract semantic features of images and improve the resolution of low-quality data. 27 The semantic features of medical images are very important for the identification of disease types. Therefore, compared with traditional algorithms of coding and filter, the generative adversarial network can process CT images better.…”
Section: Introductionmentioning
confidence: 99%