2021
DOI: 10.48550/arxiv.2106.12612
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Minimum sharpness: Scale-invariant parameter-robustness of neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 9 publications
0
2
0
Order By: Relevance
“…Related Work: Although the sharpness of the loss landscape can be calculated using several different measures (for example, the largest eigenvalue [22] or trace [23] of the Hessian of the loss), most of these measures are computationally too expensive for practical purposes. Since it has been shown that the sharpness of the loss landscape correlates with generalization [9], the main idea behind the SAM [21] algorithm is to encourage the network to seek regions where the worst loss value in a local neighbourhood is not too large (see Section 2 for more details).…”
Section: Introductionmentioning
confidence: 99%
“…Related Work: Although the sharpness of the loss landscape can be calculated using several different measures (for example, the largest eigenvalue [22] or trace [23] of the Hessian of the loss), most of these measures are computationally too expensive for practical purposes. Since it has been shown that the sharpness of the loss landscape correlates with generalization [9], the main idea behind the SAM [21] algorithm is to encourage the network to seek regions where the worst loss value in a local neighbourhood is not too large (see Section 2 for more details).…”
Section: Introductionmentioning
confidence: 99%
“…A flatter solution is a solution where the highest and the lowest loss values in the region do not differ too much. Sharpness measures in practice include the largest eigenvalue (Wu et al, 2018) or trace (Ibayashi et al, 2021) of the Hessian of the loss. Sharpness-Aware Minimization (SAM) (Foret et al, 2020) is an optimization framework that builds on the observation that sharpness of the training loss correlates with the generalization performance of a DNN.…”
Section: Introductionmentioning
confidence: 99%