2013
DOI: 10.48550/arxiv.1309.2388
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Minimizing Finite Sums with the Stochastic Average Gradient

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

5
165
1

Year Published

2015
2015
2021
2021

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 91 publications
(171 citation statements)
references
References 0 publications
5
165
1
Order By: Relevance
“…V C), as opposed to our assumption (A4) where the gradients are assumed to be known exactly. To account for noisy gradients, it is possible to use convergence bounds of stochastic gradient descent [47,58] to estimate a bound on the number of gradient descent steps. Second-order optimization methods such as stochastic reconfiguration/natural gradient [59,60] can potentially show a significant advantage over first-order optimization methods, in terms of scaling with the minimum gap of the time-dependent Hamiltonian Ĥ(t).…”
Section: (B8)mentioning
confidence: 99%
“…V C), as opposed to our assumption (A4) where the gradients are assumed to be known exactly. To account for noisy gradients, it is possible to use convergence bounds of stochastic gradient descent [47,58] to estimate a bound on the number of gradient descent steps. Second-order optimization methods such as stochastic reconfiguration/natural gradient [59,60] can potentially show a significant advantage over first-order optimization methods, in terms of scaling with the minimum gap of the time-dependent Hamiltonian Ĥ(t).…”
Section: (B8)mentioning
confidence: 99%
“…We used a variant introduced in [39], which uses Barzilai-Borwein in order to adapt the step-size, since gradient-Lipschitz constants are unavailable in the considered setting. We consider this version of variance reduction, since alternatives such as SAGA [13] and SAG [35] do not propose variants with Barzilai-Borwein type of step-size selection.…”
Section: Methodsmentioning
confidence: 99%
“…In the recent years, much effort has been made to minimize strongly convex finite sums with first order information. Recent developments, combining both numerical efficiency and sound theoretical guarantees, such as linear convergence rates, include SVRG [19], SAG [35], SDCA [36] or SAGA [13] to solve the following problem:…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The random variable Y is "realized" either by drawing an observation from an existing dataset or by using Monte Carlo. Stochastic Approximation (SA) [13], also known as Stochastic Gradient Descent (SGD), and its variants [5,7,8,11,12,14], form the popular class of methods that are used for solving problems of the type (Q). Over the past few years, there has been an increased interest in stochastic quasi-Newton methods [1,4,6,9,10,[15][16][17] that incorporate curvature information within stochastic gradient based methods.…”
Section: Introductionmentioning
confidence: 99%