2021
DOI: 10.48550/arxiv.2102.06365
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Dynamic Precision Analog Computing for Neural Networks

Abstract: Analog electronic and optical computing exhibit tremendous advantages over digital computing for accelerating deep learning when operations are executed at low precision. In this work, we derive a relationship between analog precision, which is limited by noise, and digital bit precision. We propose extending analog computing architectures to support varying levels of precision by repeating operations and averaging the result, decreasing the impact of noise. Such architectures enable programmable tradeoffs bet… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(8 citation statements)
references
References 36 publications
(75 reference statements)
0
6
0
Order By: Relevance
“…2) Post-Training Quantization: An alternative to the expensive QAT method is Post-Training Quantization (PTQ) which performs the quantization and the adjustments of the weights, without any fine-tuning [11,24,40,59,60,67,68,87,106,138,144,168,176,269]. As such, the overhead of PTQ is very low and often negligible.…”
Section: G Fine-tuning Methodsmentioning
confidence: 99%
“…2) Post-Training Quantization: An alternative to the expensive QAT method is Post-Training Quantization (PTQ) which performs the quantization and the adjustments of the weights, without any fine-tuning [11,24,40,59,60,67,68,87,106,138,144,168,176,269]. As such, the overhead of PTQ is very low and often negligible.…”
Section: G Fine-tuning Methodsmentioning
confidence: 99%
“…Quantization is one of the most widely-used techniques for neural network compression (Courbariaux et al, 2015;Han et al, 2015;Zhu et al, 2016;Zhou et al, 2016;Mishra et al, 2017;Park et al, 2017;Banner et al, 2018), with two types of training strategies: Post-Training Quantization directly quantizes a pre-trained full-precision model (He & Cheng, 2018;Nagel et al, 2019;Fang et al, 2020a;b;Garg et al, 2021); Quantization-Aware Training uses training data to optimize quantized models for better performance (Gysel et al, 2018;Esser et al, 2019;Hubara et al, 2020;Tailor et al, 2020). In this work, we focus on the latter one, which is explored in several directions.…”
Section: Related Workmentioning
confidence: 99%
“…Dynamic analog precision has been proposed by Gonugondla et al [22] to adapt to the variation in noise sensitivity across different portions of network architectures. Garg et al [24] proposed averaging the results of multiple matrix multiplications to reduce the effect of analog device noise. Our approach departs from the aforementioned proposals.…”
Section: Related Workmentioning
confidence: 99%