Our system is currently under heavy load due to increased usage. We're actively working on upgrades to improve performance. Thank you for your patience.
2020
DOI: 10.48550/arxiv.2003.00146
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

WaveQ: Gradient-Based Deep Quantization of Neural Networks through Sinusoidal Adaptive Regularization

Abstract: As deep neural networks make their ways into different domains and application, their compute efficiency is becoming a first-order constraint. Deep quantization, which reduces the bitwidth of the operations (below eight bits), offers a unique opportunity as it can reduce both the storage and compute requirements of the network superlinearly. However, if not employed with diligence, this can lead to significant accuracy loss. Due to the strong inter-dependence between layers and exhibiting different characteris… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 9 publications
0
2
0
Order By: Relevance
“…Executing machine learning algorithms in reduced precision, i.e. quantization [30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45], and through compression, i.e. sparsity/pruning [46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64], enables us to optimize the NN significantly.…”
Section: Reduced Precision and Compressionmentioning
confidence: 99%
“…Executing machine learning algorithms in reduced precision, i.e. quantization [30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45], and through compression, i.e. sparsity/pruning [46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64], enables us to optimize the NN significantly.…”
Section: Reduced Precision and Compressionmentioning
confidence: 99%
“…Pouransari et al [31] proposed the least squares quantization method that searches proper scale factors for binary quantization. Elthakeb et al [32] attempted to quantize weights and activations by applying sinusoidal regularization. The regularization involves two hyperparameters that determine weight-quantization and bitwidth-regularization strengths.…”
Section: Related Workmentioning
confidence: 99%