2018
DOI: 10.48550/arxiv.1807.06752
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Gradient Band-based Adversarial Training for Generalized Attack Immunity of A3C Path Finding

Abstract: As adversarial attacks pose a serious threat to the security of AI system in practice, such attacks have been extensively studied in the context of computer vision applications. However, few attentions have been paid to the adversarial research on automatic path finding. In this paper, we show dominant adversarial examples are effective when targeting A3C path finding, and design a Common Dominant Adversarial Examples Generation Method (CDG) to generate dominant adversarial examples against any given map. In a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0
1

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(15 citation statements)
references
References 2 publications
0
14
0
1
Order By: Relevance
“…In another related work, Bai et al [217] also attacked the Deep Q Network (DQN) [218] for robotic path-finding in a white-box setup. Similarly, Chen et al [219] also explored adversarial attacks for the same problem, and devised a so-called Common Dominant Adversarial Examples Generation Method for computing adversarial examples for a given map. In light of their threat models, we can categorize [216], [217] and [219] as white-box attacks within the RL context.…”
Section: B Reinforcement Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…In another related work, Bai et al [217] also attacked the Deep Q Network (DQN) [218] for robotic path-finding in a white-box setup. Similarly, Chen et al [219] also explored adversarial attacks for the same problem, and devised a so-called Common Dominant Adversarial Examples Generation Method for computing adversarial examples for a given map. In light of their threat models, we can categorize [216], [217] and [219] as white-box attacks within the RL context.…”
Section: B Reinforcement Learningmentioning
confidence: 99%
“…Similarly, Chen et al [219] also explored adversarial attacks for the same problem, and devised a so-called Common Dominant Adversarial Examples Generation Method for computing adversarial examples for a given map. In light of their threat models, we can categorize [216], [217] and [219] as white-box attacks within the RL context. We can also find early instances of black-box attacks for RL.…”
Section: B Reinforcement Learningmentioning
confidence: 99%
“…Figure 14 depicts a basic depiction of how an adversary can compromise the integrity of the DRL process by adding perturbations to the environment. Chen et al [63] propose a common dominant adversarial examples generation method (CDG) for crafting adversarial examples with high confidence for the environment of DRL. The proposed adversarial attack is tested on A3C trained for pathfinding.…”
Section: Attacks Targeting the Environmentmentioning
confidence: 99%
“…They argue that this resilience in NoisyNets is due to the enhanced generalize-ability and reduced transferability. Chen et al [63] propose a gradient-based adversarial training technique. They use adversarial perturbations generated using their proposed attacking algorithm, i.e., CDG, for re-training the RL agent.…”
Section: A Adversarial Trainingmentioning
confidence: 99%
“…The following question thus arises: is there similar adversarial examples that may attack the DRL algorithms in the domain of robot path planning? Some prior work has found the adversarial examples that exist in the process of robot path planning using some DRL algorithms, such as DQN, A3C, and VIN [1,3,10]. From the literature review, we can find some commonalities between the adversarial examples.…”
Section: Introductionmentioning
confidence: 98%