2020
DOI: 10.48550/arxiv.2002.07686
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Robust Quantization: One Model to Rule Them All

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 0 publications
0
5
0
Order By: Relevance
“…This fact, that in theory is not guaranteed, in practice is well established especially for the models trained from scratch, where the initial parameters distributions are typically Normal-like and surely well verified for the models considered in the Intel-Habana Labs paper, which are mostly based on convolutional architectures. Their paper [1] concludes with the statement:…”
Section: A Very Smart Idea From Our Colleagues In Intel-habana Labsmentioning
confidence: 94%
See 2 more Smart Citations
“…This fact, that in theory is not guaranteed, in practice is well established especially for the models trained from scratch, where the initial parameters distributions are typically Normal-like and surely well verified for the models considered in the Intel-Habana Labs paper, which are mostly based on convolutional architectures. Their paper [1] concludes with the statement:…”
Section: A Very Smart Idea From Our Colleagues In Intel-habana Labsmentioning
confidence: 94%
“…Intel-Habana Labs have recently published a paper [1] which considers the problem of steering the training procedure in such a way that the training itself is guided to finding solutions (that is, a set of parameters in parameter space) such that the subsequent quantization of the model is minimally harmful. Their paper considers the case of uniform symmetrical quantization and shows that in such settings, among all possible distributions, the uniform distribution is the one that yields the minimal quantization sensitivity (theorem 4 in the paper).…”
Section: A Very Smart Idea From Our Colleagues In Intel-habana Labsmentioning
confidence: 99%
See 1 more Smart Citation
“…Later, [43] finds that quantized DNNs are actually more vulnerable to adversarial attacks due to the error amplification effect, i.e., the magnitude of adversarial perturbation is amplified when passing through the DNN layers. To tackle this effect, [43,68] propose robustness-aware regularization methods for DNN training, and [69] retrains the network via feedback learning [70]. In addition, [55] searches for layerwise precision and [26] constructs a unified formulation to balance and enforce the models' robustness and compactness.…”
Section: Related Workmentioning
confidence: 99%
“…NB-SMT mitigates hardware underutiliation of DNNs caused by unstructured sparsity. In contrast to GPPs, which are obliged to keep program semantics so as to produce consistent results, DNNs are able to cope with, for example, pruning of activations and weights [5,20] and with quantization [1,17] without retraining. NB-SMT exploits DNN resiliency to avoid backpressure [19] by "squeezing" more than one thread into the shared resource by temporarily reducing the threads' numerical precision.…”
Section: Introductionmentioning
confidence: 99%