2019
DOI: 10.48550/arxiv.1906.03193
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Fighting Quantization Bias With Bias

Abstract: Low-precision representation of deep neural networks (DNNs) is critical for efficient deployment of deep learning application on embedded platforms, however, converting the network to low precision degrades its performance. Crucially, networks that are designed for embedded applications usually suffer from increased degradation since they have less redundancy. This is most evident for the ubiquitous MobileNet architecture [10,20] which requires a costly quantization-aware training cycle to achieve acceptable p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

1
17
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(18 citation statements)
references
References 18 publications
1
17
0
Order By: Relevance
“…Bias Correction. Quantization of weights induce bias shifts to activation means that may lead to detrimental behaviour in the following layers [23,32]. can be expressed as follows:…”
Section: Weight Quantizationmentioning
confidence: 99%
See 1 more Smart Citation
“…Bias Correction. Quantization of weights induce bias shifts to activation means that may lead to detrimental behaviour in the following layers [23,32]. can be expressed as follows:…”
Section: Weight Quantizationmentioning
confidence: 99%
“…Several works propose approaches to correct the quantization induced bias. These include using batch-normalization statistics [23], micro training [32] and applying scale and shift per channel [33].…”
Section: Weight Quantizationmentioning
confidence: 99%
“…Unfortunately, because the task loss function of DNNs is non-convex (Goodfellow et al, 2016), there is no analytical solution to find the form of post-training quantized weights. As a result, various approximations are being suggested mainly by using quadratic approximations that imply quantization is performed in a convex-like regime (Nagel et al, 2017;Nahshan et al, 2020).…”
Section: Weight Quantization Strategymentioning
confidence: 99%
“…Bias correction is an operation to compensate for the biased error in output activations after quantization. The amount of shift induced by quantization is diminished by adjusting the bias parameters of the neurons or channels because shifted output activations through quantization may degrade the quantization quality of the next layer (Finkelstein et al, 2019;Nagel et al, 2019). The amount of shift can be calculated as the expected error on the output activations that can be expressed as γ n = -0.9…”
Section: Bias Correction Of Q-ratermentioning
confidence: 99%
See 1 more Smart Citation