2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00958
|View full text |Cite
|
Sign up to set email alerts
|

NISP: Pruning Networks Using Neuron Importance Score Propagation

Abstract: To reduce the significant redundancy in deep Convolutional Neural Networks (CNNs), most existing methods prune neurons by only considering statistics of an individual layer or two consecutive layers (e.g., prune one layer to minimize the reconstruction error of the next layer), ignoring the effect of error propagation in deep networks. In contrast, we argue that it is essential to prune neurons in the entire neuron network jointly based on a unified goal: minimizing the reconstruction error of important respon… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
410
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 669 publications
(443 citation statements)
references
References 32 publications
2
410
0
Order By: Relevance
“…Luo et al [26] propose to use a greedy per-layer procedure to find the subset of neurons that minimize a re-construction loss, at a significant computational cost. Yu et al [34] estimate the importance of input features to a linear classifier and propagate their importance assuming Lipschitz continuity, requiring additional computational costs and nontrivial implementation of the feature score computation. Our proposed method is able to outperform these methods while requiring little additional computation and engineering.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Luo et al [26] propose to use a greedy per-layer procedure to find the subset of neurons that minimize a re-construction loss, at a significant computational cost. Yu et al [34] estimate the importance of input features to a linear classifier and propagate their importance assuming Lipschitz continuity, requiring additional computational costs and nontrivial implementation of the feature score computation. Our proposed method is able to outperform these methods while requiring little additional computation and engineering.…”
Section: Related Workmentioning
confidence: 99%
“…Pruning is a common method to derive a compact network -after training, some structural portion of the parameters is removed, along with its associated computations. A variety of pruning methods have been proposed, based on greedy algorithms [26,34], sparse regularization [20,22,33], and reinforcement learning [12]. Many of them rely on the belief that the magnitude of a weight and its importance are strongly correlated.…”
Section: Introductionmentioning
confidence: 99%
“…This is completely different from our method which aims to get the complete pruned network in one step. While some methods [23,37] also measure significance of all filters in onestep via back propagation, their estimation is biased since these approaches only depend on a mini-batch of samples.…”
Section: Discussionmentioning
confidence: 99%
“…Channel pruning was regarded as an optimization problem by Luo et al [26] and redundant channels were pruned by statistics of its next layer. Yu et al [36] conducted feature ranking to obtain neuron/channel importance score and propagated it throughout the network. The neurons/channels with smaller importance scores were removed with negligible accuracy loss.…”
Section: Related Workmentioning
confidence: 99%