2022
DOI: 10.3390/app12199453
|View full text |Cite
|
Sign up to set email alerts
|

Multiple-Stage Knowledge Distillation

Abstract: Knowledge distillation (KD) is a method in which a teacher network guides the learning of a student network, thereby resulting in an improvement in the performance of the student network. Recent research in this area has concentrated on developing effective definitions of knowledge and efficient methods of knowledge transfer while ignoring the learning ability of the student network. To fully utilize this potential learning ability and improve learning efficiency, this study proposes a multiple-stage KD (MSKD)… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 37 publications
0
1
0
Order By: Relevance
“…In the realm of multistage knowledge distillation methods, TSKD 23 effectively enhances the testing accuracy of student networks through multistage guidance from teacher networks. OtO 24 employs a joint multistage to multistage training approach between teacher and student networks, achieving significant improvements in multistage knowledge distillation.…”
Section: Related Workmentioning
confidence: 99%
“…In the realm of multistage knowledge distillation methods, TSKD 23 effectively enhances the testing accuracy of student networks through multistage guidance from teacher networks. OtO 24 employs a joint multistage to multistage training approach between teacher and student networks, achieving significant improvements in multistage knowledge distillation.…”
Section: Related Workmentioning
confidence: 99%