Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence 2018
DOI: 10.24963/ijcai.2018/330
|View full text |Cite
|
Sign up to set email alerts
|

Optimization based Layer-wise Magnitude-based Pruning for DNN Compression

Abstract: Layer-wise magnitude-based pruning (LMP) is a very popular method for deep neural network (DNN) compression. However, tuning the layer-specific thresholds is a difficult task, since the space of threshold candidates is exponentially large and the evaluation is very expensive. Previous methods are mainly by hand and require expertise. In this paper, we propose an automatic tuning approach based on optimization, named OLMP. The idea is to transform the threshold tuning problem into a constrained optimization pro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
4
1

Relationship

2
7

Authors

Journals

citations
Cited by 39 publications
(18 citation statements)
references
References 11 publications
0
18
0
Order By: Relevance
“…Both algorithms are compared on the 20 problems and the averaged final solution qualities over 20 runs are shown in Fig.5. On problems 1,4,7,8,9,12,13,14,17,18,19,20, NPDC outperforms significantly than NPDC-random. For the rest problems, NPDC only shows slightly advantages over NPDC-random.…”
Section: Investigations On the Meta-model Of Npdcmentioning
confidence: 97%
See 1 more Smart Citation
“…Both algorithms are compared on the 20 problems and the averaged final solution qualities over 20 runs are shown in Fig.5. On problems 1,4,7,8,9,12,13,14,17,18,19,20, NPDC outperforms significantly than NPDC-random. For the rest problems, NPDC only shows slightly advantages over NPDC-random.…”
Section: Investigations On the Meta-model Of Npdcmentioning
confidence: 97%
“…Evolutionary Algorithms (EAs), which work by searching the solution space of the targeted problem iteratively and in a randomized way, have shown powerful performance in solving many real-world optimization problems [5], [6], [7], [8], [9]. Unfortunately, the search-based core makes EAs ineffective and inefficient for solving large-scale optimization problems for two reasons: 1) As the number of decision variables increases, the solution space of the problem enlarges exponentially, preventing EAs exploring effectively within reasonable amount of search iterations.…”
Section: Introductionmentioning
confidence: 99%
“…[7] proposed the dynamic network surgery to recover mistaken parameter. [23] proposed an optimization algorithm to automatically tune the pruning thresholds for magnitude-based pruning methods. However, weight pruning methods always lead to unstructured models, so the model cannot leverage the existing efficient BLAS libraries in practice.…”
Section: Related Work 21 Weight Pruningmentioning
confidence: 99%
“…This method employs an optimization approach to tune the pruning threshold automatically. The pruning threshold is used to select the most important connections from all layers [20]. Another pruning approach consists in selecting the most discriminative channels based on additional discrimination-aware losses [21].…”
Section: Previous Workmentioning
confidence: 99%