2019
DOI: 10.1007/978-3-030-32430-8_14
|View full text |Cite
|
Sign up to set email alerts
|

Deceptive Reinforcement Learning Under Adversarial Manipulations on Cost Signals

Abstract: This paper studies reinforcement learning (RL) under malicious falsification on cost signals and introduces a quantitative framework of attack models to understand the vulnerabilities of RL. Focusing on Q-learning, we show that Q-learning algorithms converge under stealthy attacks and bounded falsifications on cost signals. We characterize the relation between the falsified cost and the Q-factors as well as the policy learned by the learning agent which provides fundamental limits for feasible offensive and de… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
45
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 49 publications
(45 citation statements)
references
References 26 publications
0
45
0
Order By: Relevance
“…Drones equipped with RL techniques can be commanded to collide to a crowd or a building. 32,33 Everitt et al 33 and Wang et al 34 investigated RL algorithms under corrupted reward signals. Lin et al 35 and Behzadan and Munir 36 focused on deep RL which involves DNNs for function approximation.…”
Section: Rl Securitymentioning
confidence: 99%
See 2 more Smart Citations
“…Drones equipped with RL techniques can be commanded to collide to a crowd or a building. 32,33 Everitt et al 33 and Wang et al 34 investigated RL algorithms under corrupted reward signals. Lin et al 35 and Behzadan and Munir 36 focused on deep RL which involves DNNs for function approximation.…”
Section: Rl Securitymentioning
confidence: 99%
“…Behzadan and Munir et al 31 discovered that the self‐driving platooning vehicles can collide with each other when their observation data are manipulated. Drones equipped with RL techniques can be commanded to collide to a crowd or a building 32,33 33 and Wang et al 34 investigated RL algorithms under corrupted reward signals.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In the context poisoning attacks, the adversary can modify the context observed by the agent without changing the reward associated with the context. There are also some recent interesting work on adversarial attacks against reinforcement learning algorithms under various setting [18,19,20,21,22,23,24,25].…”
Section: Introductionmentioning
confidence: 99%
“…While there is much existing work addressing adversarial attacks on supervised learning models [Szegedy et al, 2014, Goodfellow et al, 2015, Kurakin et al, 2017, Moosavi-Dezfooli et al, 2017, Wang et al, 2018, Cohen et al, 2019, Dohmatob, 2019, Wang et al, 2019, Carmon et al, 2019, Pinot et al, 2019, Alayrac et al, 2019, Dasgupta et al, 2019, Cicalese et al, 2020, Li et al, 2021, the understanding of adversarial attacks on RL models is less complete. Among the limited existing works on adversarial attacks against RL, they formally or experimentally considers different types of poisoning attack [Huang and Zhu, 2019, Sun et al, 2021, Rakhsha et al, 2020, 2021b. [Sun et al, 2021] discusses the differences between the poisoning attacks.…”
Section: Introductionmentioning
confidence: 99%