2021
DOI: 10.48550/arxiv.2101.00135
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Faults in Deep Reinforcement Learning Programs: A Taxonomy and A Detection Approach

Abstract: A growing demand is witnessed in both industry and academia for employing Deep Learning (DL) in various domains to solve real-world problems. Deep Reinforcement Learning (DRL) is the application of DL in the domain of Reinforcement Learning (RL). Like any software systems, DRL applications can fail because of faults in their programs. In this paper, we present the first attempt to categorize faults occurring in DRL programs. We manually analyzed 761 artifacts of DRL programs (from Stack Overflow posts and GitH… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 16 publications
(29 reference statements)
0
1
0
Order By: Relevance
“…As an example, Denchmark [37] does not provide enough information to reproduce and trigger bugs. Meanwhile, several studies on the testing of ML-based systems have used synthetic bugs for assessment [54,52] which may bias their evaluation by hiding potential weaknesses. Some others have also used a limited number of real bugs [79,62] that may not be representative of a thorough evaluation, implying an incorrect measure of the proposed approach's reliability.…”
Section: Introductionmentioning
confidence: 99%
“…As an example, Denchmark [37] does not provide enough information to reproduce and trigger bugs. Meanwhile, several studies on the testing of ML-based systems have used synthetic bugs for assessment [54,52] which may bias their evaluation by hiding potential weaknesses. Some others have also used a limited number of real bugs [79,62] that may not be representative of a thorough evaluation, implying an incorrect measure of the proposed approach's reliability.…”
Section: Introductionmentioning
confidence: 99%