2021
DOI: 10.48550/arxiv.2110.07810
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Towards Statistical and Computational Complexities of Polyak Step Size Gradient Descent

Abstract: We study the statistical and computational complexities of the Polyak step size gradient descent algorithm under generalized smoothness and Lojasiewicz conditions of the population loss function, namely, the limit of the empirical loss function when the sample size goes to infinity, and the stability between the gradients of the empirical and population loss functions, namely, the polynomial growth on the concentration bound between the gradients of sample and population loss functions. We demonstrate that the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(8 citation statements)
references
References 26 publications
0
8
0
Order By: Relevance
“…The proof idea of Proposition 1 follows the proof argument from [28] and subsequently Lemma 4 in Appendix E of [33]; therefore, the proof is omitted.…”
Section: Optimization Rate For the Egd Algorithmmentioning
confidence: 99%
See 2 more Smart Citations
“…The proof idea of Proposition 1 follows the proof argument from [28] and subsequently Lemma 4 in Appendix E of [33]; therefore, the proof is omitted.…”
Section: Optimization Rate For the Egd Algorithmmentioning
confidence: 99%
“…Under the high SNR regime, the population least-square function L is locally strongly convex and smooth (see Section 3.1 in [33]). Furthermore, there exist universal constants C 1 and C 2 such that with probability 1 − δ…”
Section: Generalized Linear Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition, our methodology focuses on the situation when the loss function is globally or locally strong convex. How to generalize our theoretical results to more general locally convex settings (Ho et al, 2020;Ren et al, 2022) remains to be a challenging but also an interesting topic for future study.…”
Section: Discussionmentioning
confidence: 99%
“…It is remarkable that Corollary 1 assumed locally strong convexity. To deal with more general locally convex settings, a novel method has been developed by Ho et al (2020) and Ren et al (2022) for studying the estimation error and the computational complexity of algorithm-based estimators. The key idea is to construct an interesting reference estimator sequence for the actual algorithm-based estimator sequence.…”
Section: General Loss Functionsmentioning
confidence: 99%