Proceedings of the 29th ACM International Conference on Multimedia 2021
DOI: 10.1145/3474085.3475353
|View full text |Cite
|
Sign up to set email alerts
|

Hybrid Network Compression via Meta-Learning

Abstract: Neural network pruning and quantization are two major lines of network compression. This raises a natural question that whether we can find the optimal compression by considering multiple network compression criteria in a unified framework. This paper incorporates two criteria and seeks layer-wise compression by leveraging the meta-learning framework. A regularization loss is applied to unify the constraint of input and output channel numbers, bit-width of network activations and weights, so that the compresse… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 25 publications
0
3
0
Order By: Relevance
“…In Ref. [123], the authors have jointly considered network pruning and quantization in an end-to-end meta-learning framework. Other papers, e.g.…”
Section: Conclusion Open Problems and Research Directionsmentioning
confidence: 99%
“…In Ref. [123], the authors have jointly considered network pruning and quantization in an end-to-end meta-learning framework. Other papers, e.g.…”
Section: Conclusion Open Problems and Research Directionsmentioning
confidence: 99%
“…By consolidating both meta-learning and model compression, existing researches learn light-weight models that can quickly adapt to edge environments [263]- [266]. [264] proposes an end-to-end framework to seek layer-wise compression with meta-learning. [265] learns a meta-teacher that is generalizable across domains and guides the student model to solve domain-specific tasks.…”
Section: Meta-learningmentioning
confidence: 99%
“…Through parameter quantization, the resolution of the parameters can be reduced to 16-bit, 8-bit, 4-bit, and even 1-bit with little loss of the accuracy in some tasks. [26][27][28] Prakosa et al 29 adopted the K-D method to improve the performance of the pruned network. Blakeney et al 30 proposed a parallel block-wise K-D method to compress the deep neural networks.…”
Section: Introductionmentioning
confidence: 99%