The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2019
DOI: 10.48550/arxiv.1906.04721
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Data-Free Quantization Through Weight Equalization and Bias Correction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
34
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 18 publications
(34 citation statements)
references
References 0 publications
0
34
0
Order By: Relevance
“…Minimizing KL divergence between full-precision weight distribution and quantized weight distribution is also proposed (Migacz, 2017). Since input data is not utilized, the quantization process can be simple and fast (Nagel et al, 2019) even though the correlation between weight quantization and task loss is not deeply investigated.…”
Section: Weight Quantization Strategymentioning
confidence: 99%
See 2 more Smart Citations
“…Minimizing KL divergence between full-precision weight distribution and quantized weight distribution is also proposed (Migacz, 2017). Since input data is not utilized, the quantization process can be simple and fast (Nagel et al, 2019) even though the correlation between weight quantization and task loss is not deeply investigated.…”
Section: Weight Quantization Strategymentioning
confidence: 99%
“…Bias correction is an operation to compensate for the biased error in output activations after quantization. The amount of shift induced by quantization is diminished by adjusting the bias parameters of the neurons or channels because shifted output activations through quantization may degrade the quantization quality of the next layer (Finkelstein et al, 2019;Nagel et al, 2019). The amount of shift can be calculated as the expected error on the output activations that can be expressed as γ n = -0.9…”
Section: Bias Correction Of Q-ratermentioning
confidence: 99%
See 1 more Smart Citation
“…Furthermore, the optimization process is typically constrained using some prior knowledge about the input, such as high correlation between nearby pixels within an image to avoid sample over-fit. Similar approach was recently adapted for several use-cases including for the purpose of data free distillation [21,3,17] with limited success.…”
Section: Introductionmentioning
confidence: 99%
“…By approximating real-valued weights and activations using low-bit numbers, quantized neural networks (QNNs) trained with state-of-the-art algorithms (e.g., Courbariaux et al, 2015;Rastegari et al, 2016;Louizos et al, 2018;Li et al, 2019) can be shown to perform similarly as their full-precision counterparts (e.g., Jung et al, 2019;Li et al, 2019). This work focuses on the problem of post-training quantization, which aims to generate a QNN from a pretrained full-precision network, without accessing the original training data (e.g., Sung et al, 2015;Krishnamoorthi, 2018;Zhao et al, 2019;Meller et al, 2019;Banner et al, 2019;Nagel et al, 2019;Choukroun et al, 2019). This scenario appears widely in practice.…”
Section: Introductionmentioning
confidence: 99%