2021
DOI: 10.48550/arxiv.2103.10858
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Toward Compact Deep Neural Networks via Energy-Aware Pruning

Abstract: Despite of the remarkable performance, modern deep neural networks are inevitably accompanied with a significant amount of computational cost for learning and deployment, which may be incompatible with their usage on edge devices. Recent efforts to reduce these overheads involves pruning and decomposing the parameters of various layers without performance deterioration. Inspired by several decomposition studies, in this paper, we propose a novel energy-aware pruning method that quantifies the importance of eac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(7 citation statements)
references
References 41 publications
0
7
0
Order By: Relevance
“…However, these methods increasing training time by 10 times [19,20] and can be computationally expensive while optimizing extra parameters such as a soft mask particularly for large-scale networks. Other methods generate feature maps corresponding to a set of examples and then apply metrics such as rank [13], energy [14], the average percentage of zeros [21] to quantify the importance of filters, or use similarity measures such as clustering [22] on feature maps to eliminate filters corresponding to redundant feature maps. However, generating feature maps corresponding to a set of examples take extra memory resources.…”
Section: Methods To Compute Cnn Filter Importancementioning
confidence: 99%
See 1 more Smart Citation
“…However, these methods increasing training time by 10 times [19,20] and can be computationally expensive while optimizing extra parameters such as a soft mask particularly for large-scale networks. Other methods generate feature maps corresponding to a set of examples and then apply metrics such as rank [13], energy [14], the average percentage of zeros [21] to quantify the importance of filters, or use similarity measures such as clustering [22] on feature maps to eliminate filters corresponding to redundant feature maps. However, generating feature maps corresponding to a set of examples take extra memory resources.…”
Section: Methods To Compute Cnn Filter Importancementioning
confidence: 99%
“…In filter pruning, the importance of the filters is measured using either active or passive methods. Active methods [13,14] use a dataset to generate feature maps from the filters and then compute filter importance using various measures such as entropy, the average percentage of zeros on feature maps. Some active methods even identify important filters during the training of CNNs by involving extra parameters such as a soft mask for each filter, and then jointly optimising the CNN parameters and the soft mask [11,12].…”
Section: Introductionmentioning
confidence: 99%
“…Active filter pruning: Active filter pruning methods involve a dataset to compute importance of the filters. For example, Luo et al [15], Lin et al [16], and Yeom et al [21] proposed feature-map based pruning methods, where a dataset is used to produce feature maps in CNNs, and then metrics such as entropy, variance, average rank of feature maps and the average percentage of zeros are applied on the feature maps to quantify the importance of the filters.…”
Section: B Filter Pruning Methodsmentioning
confidence: 99%
“…Other methods used for comparison: We compare the proposed operator norm based pruning method with that of the entry-wise norm based methods, (a) l 1 -norm method that eliminates filters with smaller entry-wise l 1 -norm [11] and (b) geometric median (GM) method that eliminates filters with smaller l 2 -norm as measured from the geometric median of all filters [19]. We also compare the proposed pruning method with the existing active filter pruning including HRank [16] and Energy-aware pruning [21]. The HRank method opts three steps to obtain a pruned network.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation