2018
DOI: 10.1007/978-3-319-99130-6_14
|View full text |Cite
|
Sign up to set email alerts
|

Efficient On-Line Error Detection and Mitigation for Deep Neural Network Accelerators

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
22
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 33 publications
(23 citation statements)
references
References 18 publications
1
22
0
Order By: Relevance
“…In this work, we focus on the soft errors that occur in the form of bit-flips in the data path of a DNN-accelerator. Our error model is in line with the work done in [7], [16], [42], [52]. One of the major reason of soft errors in these modern hardware systems is due to the striking of high energy particles which cause the hardware to malfunction (for example bit flip) .…”
Section: A Error Injection Modelsupporting
confidence: 79%
See 1 more Smart Citation
“…In this work, we focus on the soft errors that occur in the form of bit-flips in the data path of a DNN-accelerator. Our error model is in line with the work done in [7], [16], [42], [52]. One of the major reason of soft errors in these modern hardware systems is due to the striking of high energy particles which cause the hardware to malfunction (for example bit flip) .…”
Section: A Error Injection Modelsupporting
confidence: 79%
“…Figure 1 shows an example of errors occurring in one variable with 8 bit-depth, where input (a single variable) can be represented as a sequence of target bits and, the input can be distorted at each bit level in a stochastic process. In our experiments, errors are injected in the output of every convolution layer in a DNN similar to [27], [42], [43].…”
Section: A Error Injection Modelmentioning
confidence: 99%
“…The evaluations described in Section 5 highlight the fact that each individual performance evaluation technique is limited according to a certain set of constraints and assumptions. By better understanding these, for example through the use of techniques such as sensitivity analysis of feature maps (as described in our experiment), introspection methods [21,5], fault injection [24], mutation testing [7], a combination of evidence may be found that provides a convincing argument that the performance requirements are met. Explicitly evaluating the machine learning approach and its performance evaluation measure against the set of claims defined in the assurance claim points leads to a greater level of confidence that the performance requirements have been met.…”
Section: Extrapolation Of Resultsmentioning
confidence: 99%
“…However, most of these mitigation techniques are based on redundancy, for example, DMR: dual modular redundancy [58] and TMR: triple modular redundancy [35]. The redundancy based approaches, although considered to be very effective for other application domains [19], are highly inefficient for DNN-based systems because of the compute intensive nature of the DNNs [48], and may incur significant area, power/energy, and performance overheads. Hence, a completely new set of resource-efficient reliability mechanisms is required for robust machine learning systems.…”
Section: Reliability Threatsmentioning
confidence: 99%