2017 51st Asilomar Conference on Signals, Systems, and Computers 2017
DOI: 10.1109/acssc.2017.8335699
|View full text |Cite
|
Sign up to set email alerts
|

Minimum energy quantized neural networks

Abstract: This work targets the automated minimum-energy optimization of Quantized Neural Networks (QNNs) -networks using low precision weights and activations. These networks are trained from scratch at an arbitrary fixed point precision. At iso-accuracy, QNNs using fewer bits require deeper and wider network architectures than networks using higher precision operators, while they require less complex arithmetic and less bits per weights. This fundamental trade-off is analyzed and quantified to find the minimum energy … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
69
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 123 publications
(75 citation statements)
references
References 9 publications
(14 reference statements)
0
69
0
Order By: Relevance
“…The accuracy drop is limited to 3% when running ResNet50 on Imagenet with 2-bit weights and 4-bit activations and to 6.5% when downscaling the weights and activations to 2 bits. Furthermore, the authors of [11] investigated the trade-off between energy efficiency and accuracy of QNNs, highlighting the practical effectiveness of the sub-byte fixed-point networks. At the cost of specific retraining procedures, the accuracy drop of is kept very close to the single-precision floating point counterpart while the 2 The τp thresholds absorb bias, batch normalization and the 2 −2(Q−1) factor.…”
Section: A Quantized Neural Networkmentioning
confidence: 99%
“…The accuracy drop is limited to 3% when running ResNet50 on Imagenet with 2-bit weights and 4-bit activations and to 6.5% when downscaling the weights and activations to 2 bits. Furthermore, the authors of [11] investigated the trade-off between energy efficiency and accuracy of QNNs, highlighting the practical effectiveness of the sub-byte fixed-point networks. At the cost of specific retraining procedures, the accuracy drop of is kept very close to the single-precision floating point counterpart while the 2 The τp thresholds absorb bias, batch normalization and the 2 −2(Q−1) factor.…”
Section: A Quantized Neural Networkmentioning
confidence: 99%
“…12a shows the setup for offline off-chip training. As the synaptic weights of ODIN have a 3-bit resolution, offline training is carried out with quantization-aware stochastic gradient descent (SGD) following [57], as implemented in [58] using Keras with a TensorFlow backend. The chosen optimizer is Adam, which optimizes the weights by minimizing the categorical cross-entropy loss function during several epochs, each epoch consisting in one presentation of all labeled images in the training set.…”
Section: B Neuron and Synapse Characterizationmentioning
confidence: 99%
“…Input images are split with interleaved sub-sampling into four independent 14×14 images. The sub-image pixels are converted to rate-based Poisson-distributed spike trains and sent to four one-hidden-layer fully-connected networks resulting from Adam-based quantizationaware training in Keras following [2], [3]. Layer-wise inhibitory neurons are used to compensate for rescaling of synaptic weights trained with −1 and +1 values in Keras to values of 0 and 1 in MorphIC.…”
Section: Output Classificationmentioning
confidence: 99%
“…A teacher signal is required for supervised online learning, whereas teacher-less learning is unsupervised. operations and minimizing the memory footprint, thus avoiding the high energy cost of off-chip memory accesses if all the weights can be stored into on-chip memory [2]. The accuracy drop induced by quantization can be mitigated to acceptable levels for many applications with the use of quantization-aware training techniques that propagate binary weights during the forward pass and keep full-resolution weights for backpropagation updates [3].…”
mentioning
confidence: 99%