2015
DOI: 10.48550/arxiv.1507.06149
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Data-free parameter pruning for Deep Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
87
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 71 publications
(87 citation statements)
references
References 0 publications
0
87
0
Order By: Relevance
“…Our core pruning technique is still unstructured, magnitude pruning (among many other pruning techniques, e.g., Hu et al (2016); Srinivas and Babu (2015); Dong et al (2017); Li et al (2016); Luo et al (2017); He et al (2017)). Unstructured pruning does not necessarily yield networks that execute more quickly with commodity hardware or libraries; we aim to convey insight on neural network behavior rather than suggest immediate opportunities to improve performance.…”
Section: Rewinding On Deep Network For Imagenetmentioning
confidence: 99%
“…Our core pruning technique is still unstructured, magnitude pruning (among many other pruning techniques, e.g., Hu et al (2016); Srinivas and Babu (2015); Dong et al (2017); Li et al (2016); Luo et al (2017); He et al (2017)). Unstructured pruning does not necessarily yield networks that execute more quickly with commodity hardware or libraries; we aim to convey insight on neural network behavior rather than suggest immediate opportunities to improve performance.…”
Section: Rewinding On Deep Network For Imagenetmentioning
confidence: 99%
“…Without loss of generality, we consider image classification tasks and use ResNet, as an example to discuss our proposed on-device learning solution. Image classification is important for many edge applications, and is also the target task of the related model compression and knowledge distillation works (Hinton et al, 2015;Han et al, 2015;Chen et al, 2015;Polino et al, 2018;Srinivas & Babu, 2015). ResNet is a modern architecture with streamlined convolutional layers.…”
Section: Filter Pruning Based Model Compressionmentioning
confidence: 99%
“…To deploy DNNs on resource-constrained devices, there are two general approaches. The first approach aims to compress already-trained models, using techniques such as weights sharing (Chen et al, 2015), quantization (Han et al, 2015;Kadetotad et al, 2016), and pruning (Han et al, 2015;LeCun et al, 1990;Srinivas & Babu, 2015). However, a compressed model generated by these approaches is useful only for inference; it cannot be retrained to capture user-or device-specific requirements or new data available at runtime.…”
Section: Introductionmentioning
confidence: 99%
“…In this paper, we seek to answer following questions in the context of iterative structured pruning with rewinding: (Han, Mao, and Dally 2015;Kadetotad et al 2016), knowledge distillation (Polino, Pascanu, and Alistarh 2018;Yim et al 2017), neural architecture search (Zoph and Le 2016;Pham et al 2018) and pruning (Li et al 2016;Han, Mao, and Dally 2015;Srinivas and Babu 2015;Molchanov et al 2016). There has also been substantial work in manually designing new model topology, like Mo-bileNet (Howard et al 2017) and EfficientNet (Tan and Le 2019), that are suitable for edge device deployment but are less accurate compared to traditional models like ResNet (He et al 2016).…”
Section: Our Solutionmentioning
confidence: 99%