Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence 2019
DOI: 10.24963/ijcai.2019/480
|View full text |Cite
|
Sign up to set email alerts
|

Play and Prune: Adaptive Filter Pruning for Deep Model Compression

Abstract: While convolutional neural networks (CNN) have achieved impressive performance on various classification/recognition tasks, they typically consist of a massive number of parameters. This results in significant memory requirement as well as computational overheads. Consequently, there is a growing need for filter-level pruning approaches for compressing CNN based models that not only reduce the total number of parameters but reduce the overall computation as well. We present a new min-max framework for filter-l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
26
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 54 publications
(26 citation statements)
references
References 2 publications
0
26
0
Order By: Relevance
“…Network slimming [14] imposes sparsity regularization on the scaling factors in batch normalization layers during training, so as to identify and prune insignificant channels. Unlike these previous approaches, Play and Prune [15] allows to specify the error tolerance limit instead of the pruning ratio for each layer. Wang et al [16] verify that pruning from randomly initialized weights directly can result in more diverse pruned structures with competitive performance.…”
Section: B Dnn Model Compressionmentioning
confidence: 99%
See 1 more Smart Citation
“…Network slimming [14] imposes sparsity regularization on the scaling factors in batch normalization layers during training, so as to identify and prune insignificant channels. Unlike these previous approaches, Play and Prune [15] allows to specify the error tolerance limit instead of the pruning ratio for each layer. Wang et al [16] verify that pruning from randomly initialized weights directly can result in more diverse pruned structures with competitive performance.…”
Section: B Dnn Model Compressionmentioning
confidence: 99%
“…Our choice is based on the following three reasons: (1) Other than weight pruning [7], channel pruning produces hardware-friendly models without introducing irregular sparsity [11]. (2) 1 -norm can be easily calculated for measuring the importance of filters [13], while most pruning criteria [14], [15] can only be obtained in the formal training process. (3) Unlike [7], only oneshot pruning is adopted by FL-PQSU, as the further pruning in federated training incurs additional overhead, but contributes little to performance improvement [13].…”
Section: Structured Pruningmentioning
confidence: 99%
“…Taylor [32] feature maps: |∆C(Hi,j )| = δC δHi,j Hi,j FPGM [57] weights: Wi,j * ∈ argmin j * ∈ n in ×k * ×k j ∈ [1,nout] ||x − W i,j ||2 Prune iteratively with regularization Play and Prune [56] weights: Sj = |wk| Auto-Balance [55] weights: Sj = |wk| Prune iteratively, min reconstruction error ThiNet feature maps:…”
Section: Strategymentioning
confidence: 99%
“…"Prune iteratively" is a type of pruning that is done iteratively on a trained model that alternate between pruning and fine-tuning [32]. Pruning by regularization is usually done by adding a regularization term to the original loss function in order to leave the pruning process for the optimization [55,56]. Pruning by minimizing the reconstruction error is a family of algorithms that tries to minimize the difference of outputs between the pruned and the original model.…”
Section: Strategymentioning
confidence: 99%
See 1 more Smart Citation