2017 International Joint Conference on Neural Networks (IJCNN) 2017
DOI: 10.1109/ijcnn.2017.7966324
|View full text |Cite
|
Sign up to set email alerts
|

On-chip training of recurrent neural networks with limited numerical precision

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
28
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 37 publications
(28 citation statements)
references
References 10 publications
0
28
0
Order By: Relevance
“…There is growing interest in stochastic rounding in various domains [6,7,10,[14][15][16][17][18], and this rounding mode has started appearing in hardware devices produced by Graphcore [19] and Intel [20]. In this work we proposed and compared several algorithms for simulating stochastically rounded elementary arithmetic operations via software.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…There is growing interest in stochastic rounding in various domains [6,7,10,[14][15][16][17][18], and this rounding mode has started appearing in hardware devices produced by Graphcore [19] and Intel [20]. In this work we proposed and compared several algorithms for simulating stochastically rounded elementary arithmetic operations via software.…”
Section: Discussionmentioning
confidence: 99%
“…Furthermore, stochastic rounding is being increasingly used in machine learning [13][14][15][16][17][18]. When training neural networks, in particular, it can help compensate for the loss of accuracy caused by reducing the precision at which deep neural networks are trained in fixed-point [14] as well as floating-point [17] arithmetic.…”
Section: Motivationmentioning
confidence: 99%
“…It's also a little smoother; moreover, it trains somewhat quicker than an LSTM. Due to its relative simplicity [60]. GRUs are combined into a single update gate which acts as both an i and a forget gate.…”
Section: Gated Recurrent Unit (Gru)mentioning
confidence: 99%
“…Benefiting from the good fault tolerance of neural networks [27], [28], the data format of the predictor is represented by signed integers, not by floating-point numbers. In order to study the effect of signed integers with different digits on the branch prediction accuracy, the control variable method is used to control the size of the PHT table to 512 and the length of the GHR register to 32, gradually increasing the data representation precision of signed integers, and the comparison figure shown in Figure 7 is obtained.…”
Section: Data Representation Precisionmentioning
confidence: 99%