2022
DOI: 10.48550/arxiv.2202.00992
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Tight Convergence Rate Bounds for Optimization Under Power Law Spectral Conditions

Abstract: Performance of optimization on quadratic problems sensitively depends on the low-lying part of the spectrum. For large (effectively infinitedimensional) problems, this part of the spectrum can often be naturally represented or approximated by power law distributions. In this paper we perform a systematic study of a range of classical single-step and multi-step first order optimization algorithms, with adaptive and non-adaptive, constant and non-constant learning rates: vanilla Gradient Descent, Steepest Descen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
0
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 23 publications
0
0
0
Order By: Relevance
“…Characterizations of the NTK are fundamental for this paper and given [9,11,22,34]. Convergence analysis for optimizing NTK models directly are in [65,66].…”
Section: Literature Reviewmentioning
confidence: 99%
“…Characterizations of the NTK are fundamental for this paper and given [9,11,22,34]. Convergence analysis for optimizing NTK models directly are in [65,66].…”
Section: Literature Reviewmentioning
confidence: 99%