2019 IEEE Intelligent Transportation Systems Conference (ITSC) 2019
DOI: 10.1109/itsc.2019.8917403
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Autonomy Validation in Simulation with Adaptive Stress Testing

Abstract: During the development of autonomous systems such as driverless cars, it is important to characterize the scenarios that are most likely to result in failure. Adaptive Stress Testing (AST) provides a way to search for the mostlikely failure scenario as a Markov decision process (MDP). Our previous work used a deep reinforcement learning (DRL) solver to identify likely failure scenarios. However, the solver's use of a feed-forward neural network with a discretized space of possible initial conditions poses two … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
24
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 32 publications
(24 citation statements)
references
References 14 publications
0
24
0
Order By: Relevance
“…[62] generates a scenario controlling a pedestrian to cross the road. [63] improves the last paper by using LSTM to generate initial conditions and actions in each step. Instead of defining heuristic reward functions, [64] leverage the Go-Explore framework to find failure cases.…”
Section: Adversarial Policymentioning
confidence: 99%
“…[62] generates a scenario controlling a pedestrian to cross the road. [63] improves the last paper by using LSTM to generate initial conditions and actions in each step. Instead of defining heuristic reward functions, [64] leverage the Go-Explore framework to find failure cases.…”
Section: Adversarial Policymentioning
confidence: 99%
“…Previous papers have presented solvers based on Monte Carlo tree search [10], deep reinforcement learning (DRL) [11], and go-explore [12]. In this paper, we will use DRL and the BA, with background provided for those unfamiliar with either approach.…”
Section: B Formulationmentioning
confidence: 99%
“…where Dist(s) is some measure of the simulator's closeness to a failure, and α and β scale the penalty term given when a terminal state is reached that is not a failure [21]. In practice α and β are very large to encourage the simulator to find a failure before optimizing the action sequence to reach failure.…”
Section: B Reward Functionmentioning
confidence: 99%
“…In previous work with AST, the action space has been lowdimensional [20], [21]. However, when the action space is a high-dimensional image, previous AST algorithms will not perform well.…”
Section: Action Spacementioning
confidence: 99%
See 1 more Smart Citation