2019
DOI: 10.48550/arxiv.1906.06032
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Adversarial Training Can Hurt Generalization

Abstract: While adversarial training can improve robust accuracy (against an adversary), it sometimes hurts standard accuracy (when there is no adversary). Previous work has studied this tradeoff between standard and robust accuracy, but only in the setting where no predictor performs well on both objectives in the infinite data limit. In this paper, we show that even when the optimal predictor with infinite data performs well on both objectives, a tradeoff can still manifest itself with finite data. Furthermore, since … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
67
1

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 76 publications
(74 citation statements)
references
References 12 publications
(25 reference statements)
2
67
1
Order By: Relevance
“…The AdvProp is unique in that successfully aims at performance gain for large-scale models on clean images with adversarial examples. This is different from adversarial training, which generally results in sacrificing model accuracy on the clean images to gain robustness to adversarial examples [2], [282]. There are also other instances that report performance gain on clean data by accounting for adversarial image in training.…”
Section: A Improving Model Performancementioning
confidence: 99%
“…The AdvProp is unique in that successfully aims at performance gain for large-scale models on clean images with adversarial examples. This is different from adversarial training, which generally results in sacrificing model accuracy on the clean images to gain robustness to adversarial examples [2], [282]. There are also other instances that report performance gain on clean data by accounting for adversarial image in training.…”
Section: A Improving Model Performancementioning
confidence: 99%
“…that high standard accuracy is fundamentally at odds with high robust accuracy by considering classification problems, whereas Nakkiran (2019) suggests an alternative explanation that classifiers that are simultaneously robust and accurate are complex, and may not be contained in current function classes. However, Raghunathan et al (2019) shows that the tradeoff is not due to optimization or representation issues by showing that such tradeoffs exist even for a problem with a convex loss where the optimal predictor achieves 100% standard and robust accuracy. In contrast to previous work, we provide sharp and interpretable characterizations of the robustness-accuracy tradeoffs that may arise in regression problems, albeit restricted to linear models.…”
Section: Related Workmentioning
confidence: 99%
“…However, it was soon noticed that while adversarial training could be used to improve model robustness, it often came with a corresponding decrease in accuracy on nominal (unperturbed) data. Further, various simplified theoretical models (Tsipras et al, 2018;Zhang et al, 2019;Nakkiran, 2019;Raghunathan et al, 2019;Chen et al, 2020;Javanmard et al, 2020), have been used to explain this phenomena, and to argue that such robustness-accuracy tradeoffs are unavoidable.…”
Section: Introductionmentioning
confidence: 99%
“…Previously, some works tried to analyze adversarial training from robust optimization [27], robustness generalization [21,32], training strategy [4,15,19,29,34]. However, in these works, they all focus on the averaged robustness over all classes while ignoring the possible difference among different classes.…”
Section: Introductionmentioning
confidence: 99%