Proceedings of the 2020 on Great Lakes Symposium on VLSI 2020
DOI: 10.1145/3386263.3407650
|View full text |Cite
|
Sign up to set email alerts
|

A Privacy-Preserving-Oriented DNN Pruning and Mobile Acceleration Framework

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
2

Relationship

2
6

Authors

Journals

citations
Cited by 20 publications
(9 citation statements)
references
References 15 publications
0
9
0
Order By: Relevance
“…However, these works directly apply fixed regularization terms that penalize all weights equally, incurring potential accuracy loss. Later works [21,62,81] adopt ADMM to reform the pruning problem as optimization problems with dynamic regularization penalties, thus preserving accuracy. One drawback of these methods is the requirement for the manual setting of the compression rate for each layer.…”
Section: Pruning Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…However, these works directly apply fixed regularization terms that penalize all weights equally, incurring potential accuracy loss. Later works [21,62,81] adopt ADMM to reform the pruning problem as optimization problems with dynamic regularization penalties, thus preserving accuracy. One drawback of these methods is the requirement for the manual setting of the compression rate for each layer.…”
Section: Pruning Algorithmmentioning
confidence: 99%
“…From the pruning algorithm aspect, heuristic-based pruning was first proposed in [23] and gets improvements with more sophisticated designed heuristics [19,27,36,49,74,87]. Regularizationbased pruning [21,26,39,41,43,55,56,62,69,76,77,81], on the other hand, are more mathematicsoriented. Recent works [39,51,62,81,82] achieve substantial weight reduction without hurting the accuracy by leveraging Alternating Direction Methods of Multipliers (ADMM) with dynamic regularization penalties, but these methods require the manual setting of the compression rate for each layer.…”
Section: Introductionmentioning
confidence: 99%
“…Pruning [19,71,78,79,132,134,140,171,200,265,288] Quantization [19,68,90,134,166,179,291,307,311,314] Knowledge Distillation [29,41,42,80,83,88,95,170,186,195,220,228,231,239,257,266,267,274,295,296,300,312] Low rank factorization [76,98,119,168,190,196,210,292] Conditional Computation…”
Section: Model Compressionmentioning
confidence: 99%
“…However, the compressed VGG-16 model reduced the number of convolutional layers parameters by a factor of 41.4% for CIFAR-10 and 17.5% for CIFAR-100 dataset. In [71], the authors proposed a new framework based on weight pruning and compiler optimisation for faster inference while preserving the privacy of the training dataset. This approach initially trains the DNN model as usual on the user's own data.…”
Section: Model Compressionmentioning
confidence: 99%
“…Pruning can also be adopted to reduce the model size, which determines the per-layer pruning ratio and pruning positions. With the assumption that weights with smaller magnitudes are less important for final accuracy, magnitude-based pruning [30,58,31,86,72,26,47,57,84] is widely employed to prune weights smaller than a threshold. However, the assumption is not necessarily true, and weight magnitudes can be misleading.…”
Section: Motivation and Challengesmentioning
confidence: 99%