2020 25th International Conference on Pattern Recognition (ICPR) 2021
DOI: 10.1109/icpr48806.2021.9413182
|View full text |Cite
|
Sign up to set email alerts
|

Activation Density Driven Efficient Pruning in Training

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 8 publications
0
2
0
Order By: Relevance
“…As such, many other structured pruning techniques have been suggested since [3]- [10], [18], [19], [22], [29], [31], [37]- [39]. Studies such as [14], [24], [25] leverage activations to prune the network, as is followed in our work however they all use ℓ-norm based metrics to do so.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…As such, many other structured pruning techniques have been suggested since [3]- [10], [18], [19], [22], [29], [31], [37]- [39]. Studies such as [14], [24], [25] leverage activations to prune the network, as is followed in our work however they all use ℓ-norm based metrics to do so.…”
Section: Related Workmentioning
confidence: 99%
“…Most of these scoring techniques are based on some abstraction of the way DNNs work, for example, a simple one being the smaller the ℓ-norm (or the magnitude) of a parameter the less its assumed importance, as used in studies by [8]- [12], [14], [18]- [23]. Other approaches using activations (or outputs) of neurons as an importance score such as [24]- [26] also use ℓ-norms in some way. As designing DNNs can still be considered a somewhat elusive art, it is difficult to justify these abstractions.…”
Section: Introductionmentioning
confidence: 99%