2019
DOI: 10.48550/arxiv.1909.02384
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Training High-Performance and Large-Scale Deep Neural Networks with Full 8-bit Integers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 18 publications
0
6
0
Order By: Relevance
“…DFQ unifies two efficient DNN training mindsets, i.e., dynamic selective layer update and static low-precision training, and enables a "fractional" quantization of layers during training, in contrast to either a full execution (selected) or complete non-execution (bypassed) of layers. Furthermore, DFQ introduces input-adaptive quantization at training for the first time, and automatically learns to adapt the precision of different layers' activations and gradients in contrast to current practice of low-precision training [14,22,37] that fixes layer-wise precision during training regardless of inputs.…”
Section: Design Of Pfqmentioning
confidence: 99%
See 3 more Smart Citations
“…DFQ unifies two efficient DNN training mindsets, i.e., dynamic selective layer update and static low-precision training, and enables a "fractional" quantization of layers during training, in contrast to either a full execution (selected) or complete non-execution (bypassed) of layers. Furthermore, DFQ introduces input-adaptive quantization at training for the first time, and automatically learns to adapt the precision of different layers' activations and gradients in contrast to current practice of low-precision training [14,22,37] that fixes layer-wise precision during training regardless of inputs.…”
Section: Design Of Pfqmentioning
confidence: 99%
“…We next evaluate FracTrain over three SOTA low-precision training baselines including SBM [14], DoReFa [37], and WAGEUBN [22]. Here we consider standard training settings.…”
Section: Fractrain Over Sota Low-precision Trainingmentioning
confidence: 99%
See 2 more Smart Citations
“…Recent works have explored training DNNs with reduced precisions in floating-point arithmetic domain such as bfloat16 [40], float8 [41] as well as fixed-point arithmetic domain [13], [42]. While floating-point arithmetic is not amenable to ReRam-based hardware (without modifications), the reductions in fixed-point precision can be exploited in PANTHER by reducing the MCU width (number of slices) to improve training energy and time.…”
Section: Related Workmentioning
confidence: 99%