2020
DOI: 10.1609/aaai.v34i04.5924
|View full text |Cite
|
Sign up to set email alerts
|

AutoCompress: An Automatic DNN Structured Pruning Framework for Ultra-High Compression Rates

Abstract: Structured weight pruning is a representative model compression technique of DNNs to reduce the storage and computation requirements and accelerate inference. An automatic hyperparameter determination process is necessary due to the large number of flexible hyperparameters. This work proposes AutoCompress, an automatic structured pruning framework with the following key performance improvements: (i) effectively incorporate the combination of structured pruning schemes in the automatic process; (ii) adopt the s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
86
0
1

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 147 publications
(87 citation statements)
references
References 12 publications
0
86
0
1
Order By: Relevance
“…For the ResNet-18, the proposed PKP has the highest CR and AR, reaching 85.42× and 14.70×, respectively, and the classification accuracy is decreases by 0.64% from that of the baseline. When pruning from scratch, the CR of the PKP method is 82.73×, which is significantly higher than the 54.20× CR of the method in [49]. For VGG-16 and ResNet-18, the proposed PKP has the highest CR and AR and achieves the best performance.…”
Section: Experiments and Analysismentioning
confidence: 91%
See 1 more Smart Citation
“…For the ResNet-18, the proposed PKP has the highest CR and AR, reaching 85.42× and 14.70×, respectively, and the classification accuracy is decreases by 0.64% from that of the baseline. When pruning from scratch, the CR of the PKP method is 82.73×, which is significantly higher than the 54.20× CR of the method in [49]. For VGG-16 and ResNet-18, the proposed PKP has the highest CR and AR and achieves the best performance.…”
Section: Experiments and Analysismentioning
confidence: 91%
“…Compared to the pretrained mode, the CR of the PKP is 0.72× lower, and the classification accuracy decreases by 0.31%. The CR and AR of the method [49] are 52.20× and 8.8×, respectively, when pruning from scratch, and the classification accuracy is 1.32% lower than that of the pretrained mode. Table 2 shows the kernel sparsity ratio in each layer.…”
Section: Experiments and Analysismentioning
confidence: 95%
“…Dong et al [42] proposed a search architecture called transformable architecture that combines knowledge distillation and searchability to find a good network structure. Liu et al [43] proposed a heuristic search algorithm that trains the remaining weights while pruning to obtain a structurally sparse model of weight distribution and further searches and deletes a small part of redundant weights through network structure purification. Lin et al [44] proposed a channel-pruning method based on artificial bee colony (ABC) algorithm; searching for the optimal pruning structure is regarded as an optimization problem and the ABC algorithm is integrated to solve the problem of selecting the optimal pruning structure with the best fitness automatically.…”
Section: Pruning Methodsmentioning
confidence: 99%
“…Lin et al [30] developed a method to calculate low-rank feature maps. In addition, some works [28,29] utilized the reinforcement learning algorithm to design automatic network pruning schemes.…”
Section: ) Structured Pruningmentioning
confidence: 99%