2018
DOI: 10.48550/arxiv.1806.01368
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Adversarial Reinforcement Learning Framework for Benchmarking Collision Avoidance Mechanisms in Autonomous Vehicles

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(8 citation statements)
references
References 0 publications
0
8
0
Order By: Relevance
“…The latter is concerned with multi-agent RL settings, in which agents aim to maximize their returns in competition with other agents. While some DRL security problems can be modeled as adversarial RL (e.g., [18]), this cannot be generalized as the adversary is not necessarily a learning agent.…”
Section: Threat Modelmentioning
confidence: 99%
See 3 more Smart Citations
“…The latter is concerned with multi-agent RL settings, in which agents aim to maximize their returns in competition with other agents. While some DRL security problems can be modeled as adversarial RL (e.g., [18]), this cannot be generalized as the adversary is not necessarily a learning agent.…”
Section: Threat Modelmentioning
confidence: 99%
“…An adversary may perturb the environment and its configuration by some vector δ t (i.e, s t = s t + δ t ) to manipulate the training or inference of the agent. For instance, Behzadan & Munir [18] demonstrate that through sequential reconfiguration of obstacles on a road, an adversary can manipulate the trajectory of a DRL-based autonomous vehicle at test-time.…”
Section: Drl Attack Surfacementioning
confidence: 99%
See 2 more Smart Citations
“…The results suggested this could be an effective alternative to manual testing of complex software controllers. In more recent work, Behzadan & Munir [25] demonstrated that a reinforcement learning agent could be trained to create collisions with other road vehicles, by training an agent to collide against two agents, a DNN and a rule-based system. The number of episodes to convergence and minimum time-to-collision were then used to argue the DNN was the safer control policy.…”
Section: Introductionmentioning
confidence: 99%