2019
DOI: 10.48550/arxiv.1901.11518
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Stochastic Recursive Variance-Reduced Cubic Regularization Methods

Dongruo Zhou,
Quanquan Gu

Abstract: Stochastic Variance-Reduced Cubic regularization (SVRC) algorithms have received increasing attention due to its improved gradient/Hessian complexities (i.e., number of queries to stochastic gradient/Hessian oracles) to find local minima for nonconvex finite-sum optimization. However, it is unclear whether existing SVRC algorithms can be further improved. Moreover, the semi-stochastic Hessian estimator adopted in existing SVRC algorithms prevents the use of Hessian-vector product-based fast cubic subproblem so… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(10 citation statements)
references
References 25 publications
0
10
0
Order By: Relevance
“…Note that this lower bound is restricted to deterministic algorithms, and thus does not apply to most existing algorithms for escaping saddle points as they are all randomized algorithms. For the stochastic setting with Lipschitz stochastic gradients, the best known query complexity is Õ(ǫ −3 ) [Fang et al, 2018, Zhou andGu, 2019], while the lower bound remains open. See Appendix A for further discussion of the literature.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Note that this lower bound is restricted to deterministic algorithms, and thus does not apply to most existing algorithms for escaping saddle points as they are all randomized algorithms. For the stochastic setting with Lipschitz stochastic gradients, the best known query complexity is Õ(ǫ −3 ) [Fang et al, 2018, Zhou andGu, 2019], while the lower bound remains open. See Appendix A for further discussion of the literature.…”
Section: Discussionmentioning
confidence: 99%
“…, Allen-Zhu and Li [2017] obtain similar results without the requirement of a Hessian-vector product oracle. The sharpest rates in this category have been obtained by Fang et al [2018] and Zhou and Gu [2019], who show that the iteration complexity can be further reduced to Õ(ǫ −3 ). Again, however, this line of works consists of double-loop algorithms, and it remains unclear whether they will have an impact on practice.…”
Section: Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…With a recently proposed optimal variance reduced gradient techniques applied, Spider achieves the state-of-the-art Õ( −3 ) stochastic gradient computational cost (Fang et al, 2018). 8 Very recently, Zhou & Gu (2019) and Shen et al…”
Section: Discussion On Related Workmentioning
confidence: 98%
“…There are also a number of algorithms designed for finite sum setting where f (x) = n i=1 f i (x) [Reddi et al, 2017, Allen-Zhu and Li, 2018, Fang et al, 2018, or in case when only stochastic gradients are available [Tripuraneni et al, 2018, Jin et al, 2021, including variance reduction techniques [Allen-Zhu, 2018, Fang et al, 2018]. The sharpest rates in these settings have been obtained by Fang et al [2018], Zhou and Gu [2019] and Fang et al [2019].…”
Section: Related Workmentioning
confidence: 99%