2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.00771
|View full text |Cite
|
Sign up to set email alerts
|

Relating Adversarially Robust Generalization to Flat Minima

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
20
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 25 publications
(21 citation statements)
references
References 12 publications
1
20
0
Order By: Relevance
“…Adversarial training has also been applied to corruption robustness [28,37] or to achieve robustness against multiple different attacks [39,56,60]. While [34,52] find particularly robust subsets and [57,65] employ adversarial training with weight perturbations, the poor robustness of subsets has not been addressed yet. However, we believe that subnets with poor robustness hamper the overall adversarial robustness.…”
Section: Related Workmentioning
confidence: 99%
“…Adversarial training has also been applied to corruption robustness [28,37] or to achieve robustness against multiple different attacks [39,56,60]. While [34,52] find particularly robust subsets and [57,65] employ adversarial training with weight perturbations, the poor robustness of subsets has not been addressed yet. However, we believe that subnets with poor robustness hamper the overall adversarial robustness.…”
Section: Related Workmentioning
confidence: 99%
“…Wu et al [48] empirically verified that a flatter weight loss landscape often leads to a smaller robust generalization gap in AT. Stutz et al [40] studied the relationship between robust generalization and flatness of the robust loss landscape in the weight space. However, the above analysis and defense methods are either not general or effective for improving the robust generalization.…”
Section: Robust Overfitting In Atmentioning
confidence: 99%
“…Dong et al [12] hypothesized that robust overfitting comes from the memorization effect of the model on one-hot labels: as the one-hot labels of some samples are noisy, the model will remember those "hard" samples with noisy labels, leading to a decrease in robustness. Some works studied robust overfitting with weight loss landscape [48,40]. However, there is still a lack of general understanding about the overfitting issue, and more effective mitigation solutions are urgently needed.…”
Section: Introductionmentioning
confidence: 99%
“…Recent works [5,9,65,71,75,90] suggest a promising alternative by integrating worst-case adversarial perturbations as data augmentations (i.e., AT). AT restricts the change of loss when its input is perturbed, leading to flattening the loss landscape [30,58]. As a result, the trained network's intrinsic feature manifold and loss landscape become smoother.…”
Section: Regularize Nerf With Robust Augmentationsmentioning
confidence: 99%