2020
DOI: 10.3390/math8101652
|View full text |Cite
|
Sign up to set email alerts
|

Ensemble Learning of Lightweight Deep Learning Models Using Knowledge Distillation for Image Classification

Abstract: In recent years, deep learning models have been used successfully in almost every field including both industry and academia, especially for computer vision tasks. However, these models are huge in size, with millions (and billions) of parameters, and thus cannot be deployed on the systems and devices with limited resources (e.g., embedded systems and mobile phones). To tackle this, several techniques on model compression and acceleration have been proposed. As a representative type of them, knowledge distilla… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(4 citation statements)
references
References 42 publications
(70 reference statements)
0
4
0
Order By: Relevance
“…FitNet, the first feature-based method, is introduced to align intermediate representations layer by layer between the teacher and student models, aiming to enhance the student's performance. While this approach is simple and intuitive, it may face challenges related to convergence and performance due to the lack of high-level knowledge and the capacity gap between the two networks [28], [29]. A novel Exclusivity-Consistency regularized Knowledge Distillation (EC-KD) introduces a positionaware exclusivity strategy to enhance diversity among filters within the same layer, alleviate the limitations of student models and combine weight exclusivity and feature consistency in one unified framework [30].…”
Section: A Knowledge Distillationmentioning
confidence: 99%
“…FitNet, the first feature-based method, is introduced to align intermediate representations layer by layer between the teacher and student models, aiming to enhance the student's performance. While this approach is simple and intuitive, it may face challenges related to convergence and performance due to the lack of high-level knowledge and the capacity gap between the two networks [28], [29]. A novel Exclusivity-Consistency regularized Knowledge Distillation (EC-KD) introduces a positionaware exclusivity strategy to enhance diversity among filters within the same layer, alleviate the limitations of student models and combine weight exclusivity and feature consistency in one unified framework [30].…”
Section: A Knowledge Distillationmentioning
confidence: 99%
“…ResUNet and ResUNet?? have both been used effectively for polyp segmentation in medical image analysis [40][41][42][43]. Their ability to utilize skip connections and residual learning has enabled them to handle complex and diverse image datasets effectively.…”
Section: Deep Learning Modelsmentioning
confidence: 99%
“…Deep ensemble learning models combine the advantages of deep learning and ensemble learning to improve the generalization performance of the model. In this regard, several researchers have used ensemble learning in their studies [41][42][43][44].…”
Section: Related Workmentioning
confidence: 99%