2018 IEEE High Performance Extreme Computing Conference (HPEC) 2018
DOI: 10.1109/hpec.2018.8547604
|View full text |Cite
|
Sign up to set email alerts
|

SlimNets: An Exploration of Deep Model Compression and Acceleration

Abstract: Deep neural networks have achieved increasingly accurate results on a wide variety of complex tasks. However, much of this improvement is due to the growing use and availability of computational resources (e.g use of GPUs, more layers, more parameters, etc). Most state-of-the-art deep networks, despite performing well, over-parameterize approximate functions and take a significant amount of time to train. With increased focus on deploying deep neural networks on resource constrained devices like smart phones, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 10 publications
(9 citation statements)
references
References 3 publications
(3 reference statements)
0
9
0
Order By: Relevance
“…Representative work include the SqueezeNet [243] and the MobileNet [240]. Another method is compressing an existing network to decrease the number of parameters and the required computation resource, under the guarantee of reconstruction accuracy [244,245,246,247,248]. For example, the model cutting method compresses the model by cutting unimportant connections of a trained model according to some effective evaluations [249].…”
Section: Designing Light and Efficient Architecturesmentioning
confidence: 99%
“…Representative work include the SqueezeNet [243] and the MobileNet [240]. Another method is compressing an existing network to decrease the number of parameters and the required computation resource, under the guarantee of reconstruction accuracy [244,245,246,247,248]. For example, the model cutting method compresses the model by cutting unimportant connections of a trained model according to some effective evaluations [249].…”
Section: Designing Light and Efficient Architecturesmentioning
confidence: 99%
“…Teacher-Student training paradigm: Knowledge Distillation (KD) developed by Hinton et al (2015) is a popular technique to compress deep and wide networks into sparser ones, where the compressed model mimics the distribution learned by the complex model. Oguntola et al (2018) show that compression techniques such as pruning, and low-rank decomposition can be combined with KD to significantly improve compression rate, while maintaining accuracy. KD usually optimizes a weighted average of two different objective functions.…”
Section: Related Workmentioning
confidence: 99%
“…Pruning+KD s combines pruning with knowledge distillation using KL-divergence as described in SlimNets (Oguntola et al, 2018). Pruning+KD o applies pruning together with our version of knowledge distillation (KD o ) that combines KL-divergence with MSE-loss.…”
Section: Techniques In Comparisonmentioning
confidence: 99%
“…The success of deep convolutional neural networks (CNNs) has been well demonstrated in several real-world applications, e.g. , image classification [20,29], object detection [35], semantic segmentation [28], and low-level computer vision [37]. Massive parameters and huge computational complexity are usually required for achieving the desired high performance, which limits the application of these models to portable devices such as mobile phones and smart cameras.…”
Section: Introductionmentioning
confidence: 99%