2017
DOI: 10.48550/arxiv.1701.04143
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Vulnerability of Deep Reinforcement Learning to Policy Induction Attacks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
16
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
2
2
1

Relationship

1
4

Authors

Journals

citations
Cited by 9 publications
(17 citation statements)
references
References 0 publications
0
16
0
Order By: Relevance
“…This suggests that transferability is not an inherent property of non-robust ML models and that it may be possible to prevent black-box attacks (at least for some model classes) despite the existence of adversarial examples. 3…”
Section: Limits Of Transferabilitymentioning
confidence: 99%
See 2 more Smart Citations
“…This suggests that transferability is not an inherent property of non-robust ML models and that it may be possible to prevent black-box attacks (at least for some model classes) despite the existence of adversarial examples. 3…”
Section: Limits Of Transferabilitymentioning
confidence: 99%
“…Through slight perturbations of a machine learning (ML) model's inputs at test time, it is possible to generate adversarial examples that cause the model to misclassify at a high rate [4,22]. Adversarial examples can be used to craft human-recognizable images that are misclassified by computer vision models [22,6,14,10,13], software containing malware but classified as benign [21,26,7,8], and game environments that force reinforcement learning agents to misbehave [9,3,12].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…While the interest in deep RL solutions is extending into numerous domains such as intelligent transportation systems [1], finance [6] and critical infrastructure [15], ensuring the security and reliability of such solutions in adversarial conditions is only at its preliminary stages. Recently, Behzadan and Munir [3] reported the vulnerability of deep reinforcement learning algorithms to both test-time and training-time attacks using adversarial examples [9]. This work was followed by a number of further investigations (e.g., [11], [12]), verifying the fragility of deep RL agents to such attacks.…”
mentioning
confidence: 98%
“…To this end, we evaluate the performance of Deep Q-Network (DQN) models trained with parameter noise, against the test-time and training-time adversarial example attacks introduced in [3]. Main contributions of this work are: The remainder of this paper is organized as follows: Section 1 reviews the relevant background of DQN, parameter noise training via the NoisyNet approach, and adversarial examples.…”
mentioning
confidence: 99%