2020
DOI: 10.1016/j.neunet.2019.12.027
|View full text |Cite
|
Sign up to set email alerts
|

Training high-performance and large-scale deep neural networks with full 8-bit integers

Abstract: Deep neural network (DNN) quantization converting floating-point (FP) data in the network to integers (INT) is an effective way to shrink the model size for memory saving and simplify the operations for compute acceleration. Recently, researches on DNN quantization develop from inference to training, laying a foundation for the online training on accelerators. However, existing schemes leaving batch normalization (BN) untouched during training are mostly incomplete quantization that still adopts high precision… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
67
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 95 publications
(67 citation statements)
references
References 23 publications
(33 reference statements)
0
67
0
Order By: Relevance
“…Table 2 presents a comparison of the accuracies with those obtained from earlier studies [13] that used 16-bit DFP and [17] that used 8-bit DFP for quantized DNN training. The proposed method achieved a smaller accuracy degradation than the that of conventional DFP8.…”
Section: Resultsmentioning
confidence: 99%
“…Table 2 presents a comparison of the accuracies with those obtained from earlier studies [13] that used 16-bit DFP and [17] that used 8-bit DFP for quantized DNN training. The proposed method achieved a smaller accuracy degradation than the that of conventional DFP8.…”
Section: Resultsmentioning
confidence: 99%
“…The significant reduction of DRAM access is the major source of Shift-BNN's high energy efficiency. As various lowerprecision training techniques [4,23,62] Scalability to larger sample size. In some high-risk applications, one may need a more robust BNN model to make decisions, thus requires training BNNs with a larger sample size to strictly approximate the loss function in Eq.1.…”
Section: Evaluation Resultsmentioning
confidence: 99%
“…Although researchers proposed many approaches to quantize weights and activations, very little work was able to quantize the gradients [16,18]. What is more, the derivatives of most quantization functions are almost everywhere zero.…”
Section: Problem Formulationmentioning
confidence: 99%