2020 International Joint Conference on Neural Networks (IJCNN) 2020
DOI: 10.1109/ijcnn48605.2020.9207235
|View full text |Cite
|
Sign up to set email alerts
|

KTAN: Knowledge Transfer Adversarial Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
12
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(12 citation statements)
references
References 14 publications
0
12
0
Order By: Relevance
“…Hint learning [24] distills a deeper and thinner student model by imitating both the soft outputs and intermediate feature representations of the teacher model. Similar works are presented in [31,19,6] but are designed mainly for classifiers.…”
Section: Knowledge Distillationmentioning
confidence: 97%
See 1 more Smart Citation
“…Hint learning [24] distills a deeper and thinner student model by imitating both the soft outputs and intermediate feature representations of the teacher model. Similar works are presented in [31,19,6] but are designed mainly for classifiers.…”
Section: Knowledge Distillationmentioning
confidence: 97%
“…The distilled knowledge is defined as soft label outputs from a large teacher network, which possibly contain the structural information among different classes. Following KD, many methods are proposed to either utilize the softmax outputs [6,19] or mimic the feature layer of the teacher network [24,30,31]. However, these methods are mainly designed for multi-label classification, which cannot adapt to object detection directly.…”
Section: Introductionmentioning
confidence: 99%
“…Fortunately, some recently proposed subsampling methods [36], [37] may be applied to eliminate these low-quality samples. Additionally, some works [38]- [41] propose to incorporate the adversarial loss of GANs into KD, but their performance is not state-of-the-art.…”
Section: Introductionmentioning
confidence: 99%
“…Although existing GAN-related KD methods [38]- [41] seem related with the proposed cGAN-KD in concept, our method is fundamentally different from them, mainly due to three reasons: (1) Our approach is the first framework that utilizes cGAN-generated samples to distill and transfer knowledge, while works [38]- [41] only incorporate adversarial losses into conventional KD methods (e.g., [4]), and they cannot achieve the state-of-the-art performance. (2) Our KD framework is applicable to both classification and regression tasks, while KD methods in [38]- [41] can only apply to classification tasks. (3) Our approach is compatible with stateof-the-art KD methods (e.g., [15], [16]) and we can generally boost their performances, while methods in [38]- [41] do not have such a merit.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation