2021
DOI: 10.48550/arxiv.2103.06231
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Quantization-Guided Training for Compact TinyML Models

Sedigh Ghamari,
Koray Ozcan,
Thu Dinh
et al.

Abstract: We propose a Quantization Guided Training (QGT) method to guide DNN training towards optimized low-bit-precision targets and reach extreme compression levels below 8-bit precision. Unlike standard quantization-aware training (QAT) approaches, QGT uses customized regularization to encourage weight values towards a distribution that maximizes accuracy while reducing quantization errors. One of the main benefits of this approach is the ability to identify compression bottlenecks. We validate QGT using state-ofthe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 25 publications
(42 reference statements)
0
1
0
Order By: Relevance
“…To overcome the severe limitations in terms of memory and battery, several approaches were proposed. Tools such as Google TensorFlow Lite 1 that these tools employ hardware-independent techniques such as compression and quantization after training [5,20].…”
Section: Energy Efficient Machine Learningmentioning
confidence: 99%
“…To overcome the severe limitations in terms of memory and battery, several approaches were proposed. Tools such as Google TensorFlow Lite 1 that these tools employ hardware-independent techniques such as compression and quantization after training [5,20].…”
Section: Energy Efficient Machine Learningmentioning
confidence: 99%