2021
DOI: 10.1016/j.neucom.2021.04.026
|View full text |Cite
|
Sign up to set email alerts
|

Interactive Knowledge Distillation for image classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(6 citation statements)
references
References 12 publications
0
6
0
Order By: Relevance
“…Knowledge distillation is another method, as used in [33], [34], which is moving knowledge from a large, complicated DNN (teacher network) model to a smaller, more straightforward DNN (student network). Although distillation has increased accuracy [35], there is a chance that information will be lost in the transfer, and training the huge model will cost in terms of computation. An alternative approach is applying transfer learning (TL) [2], [36], which utilizes model weights from previously trained models.…”
Section: Related Workmentioning
confidence: 99%
“…Knowledge distillation is another method, as used in [33], [34], which is moving knowledge from a large, complicated DNN (teacher network) model to a smaller, more straightforward DNN (student network). Although distillation has increased accuracy [35], there is a chance that information will be lost in the transfer, and training the huge model will cost in terms of computation. An alternative approach is applying transfer learning (TL) [2], [36], which utilizes model weights from previously trained models.…”
Section: Related Workmentioning
confidence: 99%
“…Knowledge distillation has been shown to significantly improve the accuracy and performance of small and simple neural networks. For example, Fu et al [28] proposes a method of knowledge distillation called interactive knowledge distillation for training a light-weight student network under the guidance of a well-trained, large teacher network that outperformed a larger and more complex DNN on several image classification tasks.…”
Section: Knowledge Distillationmentioning
confidence: 99%
“…Knowledge distillation, as used in [26], [27] involves transferring knowledge from a large and complex DNN (teacher network) to a smaller and simpler DNN (student network). While distillation has improved accuracy [28], it may result in information loss during the transfer and require additional computational cost for training the large model. Dense connection, an optimization technique used in [29], [30] connects all layers in the DNN architecture directly to each other, facilitating efficient information flow to prevent loss.…”
Section: Introductionmentioning
confidence: 99%
“…The strategy neglects to consider the previous knowledge as guidance. Most existing knowledge distillation methods transfer the prediction distribution as additional knowledge (Fu et al, 2021), while the feature relationship is not fully utilized during the knowledge distillation period. This paper proposes a novel few-shot learning algorithm based on the meta-learning and knowledge distillation strategy to improve the model's performance.…”
Section: Introductionmentioning
confidence: 99%