2020
DOI: 10.1016/j.patrec.2020.04.034
|View full text |Cite
|
Sign up to set email alerts
|

Perturbation analysis of gradient-based adversarial attacks

Abstract: After the discovery of adversarial examples and their adverse effects on deep learning models, many studies focused on finding more diverse methods to generate these carefully crafted samples. Although empirical results on the effectiveness of adversarial example generation methods against defense mechanisms are discussed in detail in the literature, an in-depth study of the theoretical properties and the perturbation effectiveness of these adversarial attacks has largely been lacking. In this paper, we invest… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 17 publications
0
4
0
Order By: Relevance
“…For all CIFAR100 and ImageNet models, only 17% and 1.4% of the entries in the confusion matrix are higher than 0 (note that for an attack that induces optimal target class diversity, 100% and 7.2% of the entries in the confusion matrix would be higher than 0, respectively). Concurrent work by [30] made a similar observation on the ImageNet dataset. They find that untargeted adversarial attacks mostly cause misclassifications into semantically similar classes.…”
Section: Distribution Of Misclassificationsmentioning
confidence: 55%
“…For all CIFAR100 and ImageNet models, only 17% and 1.4% of the entries in the confusion matrix are higher than 0 (note that for an attack that induces optimal target class diversity, 100% and 7.2% of the entries in the confusion matrix would be higher than 0, respectively). Concurrent work by [30] made a similar observation on the ImageNet dataset. They find that untargeted adversarial attacks mostly cause misclassifications into semantically similar classes.…”
Section: Distribution Of Misclassificationsmentioning
confidence: 55%
“…Artificial corruptions [36,14,37,16,11] or natural shifts [15,38] on curated data have already exposed biases and architectural vulnerabilities. Adversarial robustness [39,40,41,42,43] is a related field where models are tested against adversarial examples, which introduce imperceptible though influential perturbations on images. Contrary to such attempts, we concentrated around naturally occurring distribution shifts stemming from uncurated image data.…”
Section: Robustness Under Distribution Shiftsmentioning
confidence: 99%
“…Step approach to eliminate the iterations needed to obtain an adversarial perturbation. It improved loss function [11].…”
Section: Introductionmentioning
confidence: 96%