2016 IEEE Symposium on Security and Privacy (SP) 2016
DOI: 10.1109/sp.2016.41
|View full text |Cite
|
Sign up to set email alerts
|

Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks

Abstract: Abstract-Deep learning algorithms have been shown to perform extremely well on many classical machine learning problems. However, recent studies have shown that deep learning, like other machine learning techniques, is vulnerable to adversarial samples: inputs crafted to force a deep neural network (DNN) to provide adversary-selected outputs. Such attacks can seriously undermine the security of the system supported by the DNN, sometimes with devastating consequences. For example, autonomous vehicles can be cra… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

9
1,804
0
5

Year Published

2017
2017
2022
2022

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 2,428 publications
(1,895 citation statements)
references
References 30 publications
9
1,804
0
5
Order By: Relevance
“…Consequently, Papernot, et. al [14] proposed a technique named Defensive Distillation, which is also based on retraining the network on a dimensionally-reduced set of training data. This approach, too, was recently shown to be insufficient in mitigating adversarial examples [22].…”
Section: Performance Of Proposed Policy Induction Attackmentioning
confidence: 99%
See 1 more Smart Citation
“…Consequently, Papernot, et. al [14] proposed a technique named Defensive Distillation, which is also based on retraining the network on a dimensionally-reduced set of training data. This approach, too, was recently shown to be insufficient in mitigating adversarial examples [22].…”
Section: Performance Of Proposed Policy Induction Attackmentioning
confidence: 99%
“…[7], the results of which verify the feasibility of policy induction attacks by incurring minimal perturbations in the environment or sensory inputs of an RL system. We also discuss the insufficiency of defensive distillation [14] and adversarial training [15] techniques as state of the art countermeasures proposed against adversarial example attacks on deep learning classifiers, and present potential techniques to mitigate the effect of policy induction attacks against DQNs.…”
Section: Introductionmentioning
confidence: 99%
“…However, the underlying architecture is straightforward when it comes to facilitating the ow of information forwards and backwards, greatly alleviating the e ort in generating adversarial samples. erefore, several ideas [12,23] have been proposed to enhance the complexity of DNN models,…”
Section: Enhancing Model Complexitymentioning
confidence: 99%
“…Once the two approximating DNN models are learned, the a acker can generate adversarial samples speci c to this distillation-enhanced DNN model. Similar to [23], [12] proposed to stack an auto-encoder together with a standard DNN. It shows that this auto-encoding enhancement increases a DNN's resistance to adversarial samples.…”
Section: Input Nullificationmentioning
confidence: 99%
See 1 more Smart Citation