2021
DOI: 10.48550/arxiv.2105.12686
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Dynamic Probabilistic Pruning: A general framework for hardware-constrained pruning at different granularities

Abstract: Unstructured neural network pruning algorithms have achieved impressive compression rates. However, the resulting -typically irregular -sparse matrices hamper efficient hardware implementations, leading to additional memory usage and complex control logic that diminishes the benefits of unstructured pruning. This has spurred structured coarse-grained pruning solutions that prune entire filters or even layers, enabling efficient implementation at the expense of reduced flexibility. Here we propose a flexible ne… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
1
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 29 publications
(36 reference statements)
0
1
0
Order By: Relevance
“…Most general, hyper-parameters have been learned -rather than set heuristically -to alleviate the burden of tuning [79], [80], [81]. In case such hyper-parameters relate to the size of the model (e.g., width or depth), NAS (often called pruning in this case) is generally used with the aim to find small models that facilitate implementation in (dedicated) hardware [82], [83], [84], [85], [86], [87], [88], [89]. Moreover, models have been proposed in which the functionality is conditioned upon incoming data.…”
Section: Neural Architecture Search (Nas) and Pruningmentioning
confidence: 99%
“…Most general, hyper-parameters have been learned -rather than set heuristically -to alleviate the burden of tuning [79], [80], [81]. In case such hyper-parameters relate to the size of the model (e.g., width or depth), NAS (often called pruning in this case) is generally used with the aim to find small models that facilitate implementation in (dedicated) hardware [82], [83], [84], [85], [86], [87], [88], [89]. Moreover, models have been proposed in which the functionality is conditioned upon incoming data.…”
Section: Neural Architecture Search (Nas) and Pruningmentioning
confidence: 99%
“…In case such hyper-parameters relate to the size of the model (e.g. width or depth), NAS (often called pruning in this case) is generally used with the aim to find small models that facilitate implementation in (dedicated) hardware [79], [80], [81], [82], [83], [84], [85], [86]. Moreover, models have been proposed in which the functionality is conditioned upon incoming data.…”
Section: Neural Architecture Search (Nas) and Pruningmentioning
confidence: 99%