2018
DOI: 10.1007/978-3-319-99229-7_34
|View full text |Cite
|
Sign up to set email alerts
|

Mitigation of Policy Manipulation Attacks on Deep Q-Networks with Parameter-Space Noise

Abstract: Recent developments have established the vulnerability of deep reinforcement learning to policy manipulation attacks via intentionally perturbed inputs, known as adversarial examples. In this work, we propose a technique for mitigation of such attacks based on addition of noise to the parameter space of deep reinforcement learners during training. We experimentally verify the effect of parameter-space noise in reducing the transferability of adversarial examples, and demonstrate the promising performance of th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(4 citation statements)
references
References 21 publications
0
4
0
Order By: Relevance
“…Similarly, Pinto et al [24] introduced two agents to play the zero-sum discount game to ensure the robustness of the policy learning. Different from the adversarial playing, Behzadan et al [4] adopt an equivalent model as the noisy network to generate the adversarial samples via FGSM. Neklyudov et al [21] applied the Gaussian variance layer to generate the adversarial samples, and empirical results show that this method is effective in improving the ability of exploration and robustness of agents.…”
Section: Adversarial Defensementioning
confidence: 99%
See 1 more Smart Citation
“…Similarly, Pinto et al [24] introduced two agents to play the zero-sum discount game to ensure the robustness of the policy learning. Different from the adversarial playing, Behzadan et al [4] adopt an equivalent model as the noisy network to generate the adversarial samples via FGSM. Neklyudov et al [21] applied the Gaussian variance layer to generate the adversarial samples, and empirical results show that this method is effective in improving the ability of exploration and robustness of agents.…”
Section: Adversarial Defensementioning
confidence: 99%
“…The DRL is implemented in python based on Pytorch package 1 , and the input of Agent and the attacker from Atari's games are selected image frame which is transferred to 84×84-pixel image, and the policy network is a classical convolutional neural network which is to map the input to the action space. There are three convolutional layers with the size of (32,8,8,4), (64, 4, 4, 2) and (64, 3, 3, 1), where the first value in the bracket is the number of the filters, the second and the third values denote the filter size, and the last value represents the stride size. The last two layers are the fully connected functions that map the hidden representation to the action space, and the shape of the weights in these two layers are (3136, 512) and (512, action space).…”
Section: Settingsmentioning
confidence: 99%
“…Randomization methods [64,2] were first proposed to encourage exploration. NoisyNet [24] adds parametric noise to the network's weight during training, providing better resilience to both training-time and test-time attacks [5,6]. Under the adversarial training framework, Kos et al [38] and Behzadan et al [5] show that re-training with random noise and FGSM perturbations increases the resilience against adversarial examples.…”
Section: Related Workmentioning
confidence: 99%
“…To guarantee the security associated with the learning of RL policies, defenses against training-time attacks have been developed from the standpoint of robustness which refers to the ability of an agent to maintain its functionality in the presence of perturbations [73] (illustrated as Figure 5.1). These robustness-based defenses [61,[73][74][75][76][77][78][79][80]82] either theoretically or empirically guarantee the performance of learning policy under perturbations at training time. In spite of the fact of robustness is a crucial issue, it is merely an add-on concern when designing RL algorithms, which could increase design costs or compromise policy performance.…”
Section: Introductionmentioning
confidence: 99%