2019
DOI: 10.1186/s42400-019-0027-x
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial attack and defense in reinforcement learning-from AI security view

Abstract: Reinforcement learning is a core technology for modern artificial intelligence, and it has become a workhorse for AI applications ranging from Atrai Game to Connected and Automated Vehicle System (CAV). Therefore, a reliable RL system is the foundation for the security critical applications in AI, which has attracted a concern that is more critical than ever. However, recent studies discover that the interesting attack mode adversarial attack also be effective when targeting neural network policies in the cont… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
43
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 94 publications
(53 citation statements)
references
References 34 publications
0
43
0
Order By: Relevance
“…Other researchers identified an emerging interest for deepfake ransomware [48] in certain cybercriminal circles. Beyond that, it has been demonstrated that via a replica of a victim intelligent system (a deep reinforcement learning agent), the policies of the victim system can be compromised in a targeted way [49].…”
Section: Rda For Ai Risk Instantiations Ia and Ib-examplesmentioning
confidence: 99%
“…Other researchers identified an emerging interest for deepfake ransomware [48] in certain cybercriminal circles. Beyond that, it has been demonstrated that via a replica of a victim intelligent system (a deep reinforcement learning agent), the policies of the victim system can be compromised in a targeted way [49].…”
Section: Rda For Ai Risk Instantiations Ia and Ib-examplesmentioning
confidence: 99%
“…Would there be any guarantees for convergence in such a twisted model? Some approaches do try to create adversarial examples to make the models better suited for outliers (Pinto et al, 2017b;Chen et al, 2019). Our work is novel in that we wish for the model to learn what does not work, leading to better exploration and more accessible domain adaptation to unseen tasks not by changing the tasks or rewards but by letting the same embedding network learn both skills.…”
Section: Introductionmentioning
confidence: 99%
“…In more advanced attack models known as insider attacks, attacker falsifies the data input by considering the target DNN structure of the learning model. There are two distinct adversarial attack settings on learning agents: white-box attack where attackers have access to the training model of learning agent and interacts with target model for generating adversarial inputs, and black-box attack where malicious inputs are generated from an estimated training model which is close to the true target model of learning agent [3]. In this paper, we thoroughly investigate security vulnerabilities of DRL based TSCs under two adversarial attack models namely Fast Gradient Sign Method (FGSM) [4] and Jacobian-based Saliency Map Attack (JSMA) [5] with white-box and black-box settings.…”
Section: Introductionmentioning
confidence: 99%