The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2014
DOI: 10.48550/arxiv.1412.7024
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Training deep neural networks with low precision multiplications

Abstract: Multipliers are the most space and power-hungry arithmetic operators of the digital implementation of deep neural networks. We train a set of state-of-the-art neural networks (Maxout networks) on three benchmark datasets: MNIST, CIFAR-10 and SVHN. They are trained with three distinct formats: floating point, fixed point and dynamic fixed point. For each of those datasets and for each of those formats, we assess the impact of the precision of the multiplications on the final error after training. We find that v… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
140
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 108 publications
(141 citation statements)
references
References 14 publications
1
140
0
Order By: Relevance
“…However, postquantization yields performance loss, which is more prominent as the precision lowers. In particular, posttraining binarization (1-bit precision) enables the highest model compression and computational speedup but impacts heavily on a classifier's accuracy [27].…”
Section: B Binary Neural Networkmentioning
confidence: 99%
“…However, postquantization yields performance loss, which is more prominent as the precision lowers. In particular, posttraining binarization (1-bit precision) enables the highest model compression and computational speedup but impacts heavily on a classifier's accuracy [27].…”
Section: B Binary Neural Networkmentioning
confidence: 99%
“…Prior work has proposed schemes for uniform quantization (Courbariaux et al, 2014;Zhou et al, 2016) and nonuniform quantization (Han et al, 2015;Zhu et al, 2016). Uniform quantization uses integer or fixed-point format which can be accelerated with specialized math pipelines and is the focus of this paper.…”
Section: Related Workmentioning
confidence: 99%
“…To measure the robustness of VeriDL, we implement two types of server Model compression attack. The attack compresses a trained DNN network with small accuracy degradation [2,7]. To simulate the attack, we setup a fullyconnected network with two hidden layers and sigmoid activation function.…”
Section: Robustness Of Verificationmentioning
confidence: 99%