Proceedings of the 29th International Conference on Compiler Construction 2020
DOI: 10.1145/3377555.3377900
|View full text |Cite
|
Sign up to set email alerts
|

Robust quantization of deep neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 17 publications
0
1
0
Order By: Relevance
“…The number of accumulators (at least three) tied to each multiply and accumulate (MAC) processing element (PE) might require more hardware resource for implementation, not to mention that existing CNN accelerators usually implements large number of PE in parallel to achieve high performance computation. In the work by Kim et al [22] and Cai et al [23], they proposed efficient search algorithm to find the optimal bitwidth for each of the layer weights. This scheme results in mixed precision for the network weights, where each layer might be represented using a different bitwidth.…”
Section: B Motivation: Why Another Level Of Quantization?mentioning
confidence: 99%
“…The number of accumulators (at least three) tied to each multiply and accumulate (MAC) processing element (PE) might require more hardware resource for implementation, not to mention that existing CNN accelerators usually implements large number of PE in parallel to achieve high performance computation. In the work by Kim et al [22] and Cai et al [23], they proposed efficient search algorithm to find the optimal bitwidth for each of the layer weights. This scheme results in mixed precision for the network weights, where each layer might be represented using a different bitwidth.…”
Section: B Motivation: Why Another Level Of Quantization?mentioning
confidence: 99%