2018
DOI: 10.1007/978-3-030-01237-3_12
|View full text |Cite
|
Sign up to set email alerts
|

A Systematic DNN Weight Pruning Framework Using Alternating Direction Method of Multipliers

Abstract: Weight pruning methods for deep neural networks (DNNs) have been investigated recently, but prior work in this area is mainly heuristic, iterative pruning, thereby lacking guarantees on the weight reduction ratio and convergence time. To mitigate these limitations, we present a systematic weight pruning framework of DNNs using the alternating direction method of multipliers (ADMM). We first formulate the weight pruning problem of DNNs as a nonconvex optimization problem with combinatorial constraints specifyin… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
228
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3

Relationship

3
4

Authors

Journals

citations
Cited by 382 publications
(244 citation statements)
references
References 27 publications
0
228
0
Order By: Relevance
“…Problem Formulation: Consider an N -layer DNN, and we focus on the most computationally intensive CONV layers. The weights and biases of layer k are respectively denoted by W k and b k , and the loss function of DNN is denoted by [64] for more details. In our discussion, {W k } N k=1 and {b k } N k =1 respectively characterize the collection of weights and biases from layer 1 to layer N .…”
Section: Kernel Pattern and Connectivity Pruning Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…Problem Formulation: Consider an N -layer DNN, and we focus on the most computationally intensive CONV layers. The weights and biases of layer k are respectively denoted by W k and b k , and the loss function of DNN is denoted by [64] for more details. In our discussion, {W k } N k=1 and {b k } N k =1 respectively characterize the collection of weights and biases from layer 1 to layer N .…”
Section: Kernel Pattern and Connectivity Pruning Algorithmmentioning
confidence: 99%
“…Early efforts on DNN model compression [8,12,14,15,19,42,54] mainly rely on iterative and heuristic methods, with limited and non-uniform model compression rates. Recently, a systematic DNN model compression framework (ADMM-NN) has been developed using the powerful mathematical optimization tool ADMM (Alternating Direction Methods of Multipliers) [4,21,39], currently achieving the best performance (in terms of model compression rate under the same accuracy) on weight pruning [49,64] and one of the best on weight quantization [35].…”
Section: Introductionmentioning
confidence: 99%
“…FPGA hardware accelerators [19], [20] have also been investigated to accommodate pruned CNNs, by leveraging the reconfigurability in on-chip resources. Recently, the authors of [14] have developed a systematic weight pruning framework based on the powerful optimization tool ADMM (Alternating Direction Method of Multipliers) [21]. Such framework consistently achieves higher pruning rate than prior arts.…”
Section: B Cnn Weight Pruningmentioning
confidence: 99%
“…Different from the existing ADMM based approach [14] which performs pruning in the space domain, SPEC 2 performs end-to-end spectral pruning without transformation between W and W . Such spectral pruning enables us to exploit computation redundancy in both the sliding window operation and the spectral kernel weights.…”
Section: A Overviewmentioning
confidence: 99%
“…In order to deploy DNNs on these embedded devices, DNN model compression techniques such as weight pruning, have been proposed for storage reduction and computation acceleration. Recently, works such as [5,20] have made breakthrough on the weight pruning methods for DNNs while maintaining the network accuracy. However, the network structure and weight storage after pruning become highly irregular and therefore the storage of indexing is non-negligible, which undermines the compression ratio and the performance.…”
Section: Introductionmentioning
confidence: 99%