Proceedings of the 29th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Softw 2021
DOI: 10.1145/3468264.3468612
|View full text |Cite
|
Sign up to set email alerts
|

Exposing numerical bugs in deep learning via gradient back-propagation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

3
5

Authors

Journals

citations
Cited by 30 publications
(13 citation statements)
references
References 36 publications
1
12
0
Order By: Relevance
“…The Training stage tends to be time-consuming due to heavy numerical computation based on a large amount of training data. As presented in the existing study [51], the typical training time ranges from a few minutes to several days. Hence, the bugs observed at this stage, especially those Incorrect Functionality bugs (account for 30.7% of bugs observed at this stage), may be manifested after hours or even days into the training process.…”
Section: Symptom Distributionmentioning
confidence: 99%
See 2 more Smart Citations
“…The Training stage tends to be time-consuming due to heavy numerical computation based on a large amount of training data. As presented in the existing study [51], the typical training time ranges from a few minutes to several days. Hence, the bugs observed at this stage, especially those Incorrect Functionality bugs (account for 30.7% of bugs observed at this stage), may be manifested after hours or even days into the training process.…”
Section: Symptom Distributionmentioning
confidence: 99%
“…A DL system typically involves three levels [51]: the production level (i.e., DL models), program level (i.e., DL programs used for training DL models), and framework level (i.e., DL frameworks used by developers for implementing DL programs). Bugs in any level could affect the overall quality of the DL system.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…For example, when generating 20-operator graphs, FP exceptional values occur in 56.8% of generated graphs if we use random weight and inputs. This is because some operators, which we refer to as vulnerable operators [63], produce real (e.g., √ 𝑥 returns 𝑁 𝑎𝑁 if 𝑥 < 0) or stable (e.g., 𝑥 𝑦 returns 𝐼𝑛𝑓 for large 𝑥 and 𝑦) results only for a subset of their input domain. If a vulnerable operator's input lies outside of this domain, the operator outputs an FP exceptional value, which propagates through the model and impacts the model's output, preventing us from comparing model outputs during differential testing.…”
Section: Improving Numeric Validity With Gradientsmentioning
confidence: 99%
“…Software testing is one of the most important parts of the entire cycle of software development and is vital to guarantee the quality of software [20,47,60,61]. As a kind of important software, similar to the quality assurance of other software, compiler testing is also one of the most widely-used ways of guaranteeing the quality of compilers and has received extensive attention from both practitioners and researchers in the area of software engineering [8, 19, 22ś24, 36, 56].…”
Section: Introductionmentioning
confidence: 99%