2019 IEEE/ACM International Conference on Computer-Aided Design (ICCAD) 2019
DOI: 10.1109/iccad45719.2019.8942068
|View full text |Cite
|
Sign up to set email alerts
|

ALWANN: Automatic Layer-Wise Approximation of Deep Neural Network Accelerators without Retraining

Abstract: The state-of-the-art approaches employ approximate computing to reduce the energy consumption of DNN hardware. Approximate DNNs then require extensive retraining afterwards to recover from the accuracy loss caused by the use of approximate operations. However, retraining of complex DNNs does not scale well. In this paper, we demonstrate that efficient approximations can be introduced into the computational path of DNN accelerators while retraining can completely be avoided. ALWANN provides highly optimized imp… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
113
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 93 publications
(113 citation statements)
references
References 28 publications
0
113
0
Order By: Relevance
“…We select Neural Network inference as our testcase and we use the ResNet-8 network [59] (7 convolution layers) trained with the CIFAR-10 image classification dataset. The NN is quantized at 8 bits to avoid floating point operations [60]. First, we use RETSINA to generate two approximate reconfigurable 8-bit multipliers.…”
Section: Figure 12mentioning
confidence: 99%
See 3 more Smart Citations
“…We select Neural Network inference as our testcase and we use the ResNet-8 network [59] (7 convolution layers) trained with the CIFAR-10 image classification dataset. The NN is quantized at 8 bits to avoid floating point operations [60]. First, we use RETSINA to generate two approximate reconfigurable 8-bit multipliers.…”
Section: Figure 12mentioning
confidence: 99%
“…The accuracy levels of RM1 and RM2 are selected targeting high inference accuracy. To achieve this, we used [31] and [60] to profile how the multiplier's error impacts ResNet's accuracy. Next, we extend [60] and integrate our reconfigurable multipliers RM1 and RM2.…”
Section: Figure 12mentioning
confidence: 99%
See 2 more Smart Citations
“…The works presented in [10], [13], [20], [36], [37] had used logic minimization to create the optimal approximate multipliers for each network model. Logic minimization intentionally flips bits in the logic to reduce the size of the operators, and these techniques use heuristics to find the optimal targets.…”
Section: Related Workmentioning
confidence: 99%