2020
DOI: 10.48550/arxiv.2002.02949
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Activation Density driven Energy-Efficient Pruning in Training

Abstract: The process of neural network pruning with suitable fine-tuning and retraining can yield networks with considerably fewer parameters than the original with comparable degrees of accuracy. Typically, pruning methods require large, pre-trained networks as a starting point from which they perform a timeintensive iterative pruning and retraining algorithm. We propose a novel pruning in-training method that prunes a network realtime during training, reducing the overall training time to achieve an optimal compresse… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 10 publications
(24 reference statements)
0
4
0
Order By: Relevance
“…There is relative lack of research on networks that have both forms of sparsity (exceptions are [1,94]). In some scenarios networks trained without explicit activation sparsity end up with highly sparse activations anyway [8,19,20,33]. This is encouraging because it suggests that sparse activations may naturally be an optimal outcome.…”
Section: Deploying Complex Sparse-sparse Systemsmentioning
confidence: 84%
See 3 more Smart Citations
“…There is relative lack of research on networks that have both forms of sparsity (exceptions are [1,94]). In some scenarios networks trained without explicit activation sparsity end up with highly sparse activations anyway [8,19,20,33]. This is encouraging because it suggests that sparse activations may naturally be an optimal outcome.…”
Section: Deploying Complex Sparse-sparse Systemsmentioning
confidence: 84%
“…In our implementation we use k-WTA to achieve activation sparsity. Another approach is to remove entire channels in convolutional layers during training through a structured pruning process [19,44,73]. In [19] they notice that activations naturally become sparse during training and use a measure of sparsity to gradually prune channels.…”
Section: Accelerating Sparse Network On Other Platformsmentioning
confidence: 99%
See 2 more Smart Citations