2016
DOI: 10.48550/arxiv.1608.04644
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Towards Evaluating the Robustness of Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
159
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 87 publications
(169 citation statements)
references
References 0 publications
0
159
0
Order By: Relevance
“…One major focus of these optimization problems is on testing the resilience of neural networks against adversarial attack [7]. This involves either maximizing a notion of resilience [8] or finding minimal perturbations needed to misclassify an image [16].…”
Section: Outline and Contributionsmentioning
confidence: 99%
See 1 more Smart Citation
“…One major focus of these optimization problems is on testing the resilience of neural networks against adversarial attack [7]. This involves either maximizing a notion of resilience [8] or finding minimal perturbations needed to misclassify an image [16].…”
Section: Outline and Contributionsmentioning
confidence: 99%
“…The full optimization problem is reproduced in Eqns. (7), and the associated sets are given in Table 2.…”
Section: E Rmentioning
confidence: 99%
“…• l 0 attack: OnePixel [Su et al, 2019], SparseFool [Modas et al, 2019] • l 2 attack: Projected Gradient Descent-l 2 (PGDL2) [Goodfellow et al, 2014, Madry et al, 2017, DeepFool [Moosavi-Dezfooli et al, 2015], CW attack [Carlini and Wagner, 2016], AutoAttack-l 2 [Wong et al, 2020] • l ∞ attack: Fast Gradient Sign Method (FGSM) [Goodfellow et al, 2014], Projected Gradient Descent (PGD) [Goodfellow et al, 2014, Madry et al, 2017, AutoAttack-l ∞ [Wong et al, 2020] FGSM As one of the earliest and most popular adversarial attacks described by Goodfellow et al [2014], Fast Gradient Sign Method (FGSM) serves as a baseline attack in our training. As notified previously, to optimize the parameter in trained models is to maximize the loss function over δ.…”
Section: Adversarial Attackmentioning
confidence: 99%
“…This is mainly due to the time-consuming generation of adversarial examples which alone requires an optimization procedure, e.g. via fast gradient sign method (FGSM) [Goodfellow et al, 2014], projected gradient descent (PGD) [Goodfellow et al, 2014, Madry et al, 2017, Kurakin et al, 2016, One Pixel attack [Su et al, 2019], CW attack [Carlini and Wagner, 2016], or DeepFool [Moosavi-Dezfooli et al, 2015].…”
Section: Introductionmentioning
confidence: 99%
“…Since 2014, when the observation was made that applying small perturbations to inputs can cause dramatic shifts in model outputs [16], the field of adversarial machine learning has been elevated to the forefront of computer vision research, with numerous techniques for generating so-called adversarial attacks being published each year. One such technique, FGSM [8], is based upon the model gradients, and serves as the foundation for other techniques such as PGD [13]; this class of adversarial example generation technique generally applies per-pixel perturbations budgeted according to some other techniques, such as DeepFool [14]and C&W [2], similarly apply pixel-level changes, often while simultaneously reducing the overall amount of applied perturbation and still maintaining an impressive rate of misclassifica-tion. Though many of these pixel-level perturbation techniques are generalizable from the task of image classification to broader applications, there has been some research into adapting these techniques specifically to suit the domain of facial recognition [1].…”
Section: Adversarial Attack Strategiesmentioning
confidence: 99%