2018 IEEE Third International Conference on Data Science in Cyberspace (DSC) 2018
DOI: 10.1109/dsc.2018.00126
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial Examples Construction Towards White-Box Q Table Variation in DQN Pathfinding Training

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 23 publications
(14 citation statements)
references
References 4 publications
0
14
0
Order By: Relevance
“…Other research also using DQN in optimal control has shown successful results by simulation. For instance, in [31], it exploited DQN to find the optimal path of a robotic agent in a simple 2D environment with a limited number of states and no uncertainties (a 15 × 15 grid). DQN was also used for path planning of a ground robot in the seekavoid arena 01, a virtual environment on the DeepMind Lab platform containing some visual obstacles [32].…”
Section: Discussionmentioning
confidence: 99%
“…Other research also using DQN in optimal control has shown successful results by simulation. For instance, in [31], it exploited DQN to find the optimal path of a robotic agent in a simple 2D environment with a limited number of states and no uncertainties (a 15 × 15 grid). DQN was also used for path planning of a ground robot in the seekavoid arena 01, a virtual environment on the DeepMind Lab platform containing some visual obstacles [32].…”
Section: Discussionmentioning
confidence: 99%
“…Other studies of adversarial attacks on the specific application of DRL for path-finding have also been conducted by (Xiang et al 2018) and (Bai et al 2018), which results in the RL agent failing to find a path to the goal or planning a path that is more costly.…”
Section: Adversarial Attacks On Rl Agentmentioning
confidence: 99%
“…Based on the SPA algorithm introduced above, Bai et al (2018) proposed that they first use DQN to find the optimal path, and analyzed the rules of DQN pathfinding. They proposed a method that can effectively find vulnerable points towards White-Box Q-table variation in DQN pathfinding training.…”
Section: White-box Based Adversarial Attack On Dqn (Wba)mentioning
confidence: 99%
“…FGSM (Goodfellow et al 2014a), SPA (Xiang et al 2018), WBA (Bai et al 2018), and CDG (Chen et al 2018b) belong to White-box attack, which have access to the details related to training algorithm and corresponding parameters of the target model. Meanwhile, the PIA (Behzadan and Munir 2017), STA (Lin et al 2017), EA (Lin et al 2017), and AVI (Liu et al 2017) are Black-box attacks, in which adversary has no idea of the details related to training algorithm and corresponding parameters of the model, for the threat model discussed in these literatures, authors assumed that the adversary has access to the training environment bat has no idea of the random initializations of the target policy, and additionally does not know what the learning algorithm is.…”
Section: Summary For Adversarial Attack In Reinforcement Learningmentioning
confidence: 99%
See 1 more Smart Citation