Interspeech 2022 2022
DOI: 10.21437/interspeech.2022-874
|View full text |Cite
|
Sign up to set email alerts
|

Sub-8-Bit Quantization Aware Training for 8-Bit Neural Network Accelerator with On-Device Speech Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 0 publications
0
2
0
Order By: Relevance
“…Consequently, R(W) cannot enforce each weight to be replaced by the closest centroid in z as w = arg min i ||w − z i || , for w ∈ W and z i ∈ z, because the min operator is not differentiable. Recent BP-QAT methods force weights to approach the centroid in z using R(W) = w∈W D(w, z), where the differentiable dissimilarity function D is based on a cosine function in [1,14].…”
Section: Related Qat Approachesmentioning
confidence: 99%
See 1 more Smart Citation
“…Consequently, R(W) cannot enforce each weight to be replaced by the closest centroid in z as w = arg min i ||w − z i || , for w ∈ W and z i ∈ z, because the min operator is not differentiable. Recent BP-QAT methods force weights to approach the centroid in z using R(W) = w∈W D(w, z), where the differentiable dissimilarity function D is based on a cosine function in [1,14].…”
Section: Related Qat Approachesmentioning
confidence: 99%
“…FP-QAT [11] quantizes the model weights during forward propagation to pre-defined quantization centroids. BP-QAT [1,14,15] relies on customized regularizers to gradually force weights to those quantization centroids (i.e., "soft quantization" via gradient) during training before hard compression performs in the late training phase. As model weights are informed by the customized regularizers to move closer to where they are quantized at runtime per training step, the predictive performance is often well preserved.…”
Section: Introductionmentioning
confidence: 99%