2017
DOI: 10.48550/arxiv.1703.06748
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Tactics of Adversarial Attack on Deep Reinforcement Learning Agents

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
132
1
1

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 90 publications
(143 citation statements)
references
References 0 publications
0
132
1
1
Order By: Relevance
“…Specifically, adversarial attack and defense in RS have received a lot of attention in recent years [22] as security is crucial in RS. Moreover, DRL policies are vulnerable to adversarial perturbations to agent's observations [57]. Gleave et al [30] provide an adversarial attack method for perturbing the observations, thus affecting the learned policy.…”
Section: Robustness On Adversarial Samples and Attacksmentioning
confidence: 99%
See 1 more Smart Citation
“…Specifically, adversarial attack and defense in RS have received a lot of attention in recent years [22] as security is crucial in RS. Moreover, DRL policies are vulnerable to adversarial perturbations to agent's observations [57]. Gleave et al [30] provide an adversarial attack method for perturbing the observations, thus affecting the learned policy.…”
Section: Robustness On Adversarial Samples and Attacksmentioning
confidence: 99%
“…Cao et al [10] provide an adversarial attack detection method for DRL-based RS which uses the GRU to encode the action space into a low-dimensional space and design decoders to detect the potential attack. However, it only considers Fast Gradient Sign Method (FGSM)-based attacks and strategically-timed attacks [57]. Thus, it lacks the capability to detect other types of attack.…”
Section: Robustness On Adversarial Samples and Attacksmentioning
confidence: 99%
“…In this work, we focus on the case of one-shot grid manipulation attack, and attack one specific line/bus state, where we do not need the exact state transition model. In [22], further discussion on the timing of attacks is made, while more systematic modeling of the state model s t+1 = f ((s(t), a(t)) is discussed in [23], which can craft stronger attack when the environment model can be learned by the attacker.…”
Section: B Attack Implementationmentioning
confidence: 99%
“…Generally speaking, there are two types of strategies to attack decision-making policies: (1) Attacking-on-States: directly tampering with the state sequence of the target policy. These methods [8], [9] impose small disturbances on the state observations to make the policy output wrong decisions or get a smaller reward. (2) Attacking-by-Policy: learning one or more adversarial policies through interacting with the target policy.…”
Section: B Attacks On Policiesmentioning
confidence: 99%