2019
DOI: 10.48550/arxiv.1901.08689
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Don't Jump Through Hoops and Remove Those Loops: SVRG and Katyusha are Better Without the Outer Loop

Abstract: The stochastic variance-reduced gradient method (SVRG) and its accelerated variant (Katyusha) have attracted enormous attention in the machine learning community in the last few years due to their superior theoretical properties and empirical behaviour on training supervised machine learning models via the empirical risk minimization paradigm. A key structural element in both of these methods is the inclusion of an outer loop at the beginning of which a full pass over the training data is made in order to comp… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
46
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 19 publications
(47 citation statements)
references
References 12 publications
1
46
0
Order By: Relevance
“…Note that those findings are in accord with Corollary H.3. Similar results were shown in [15] for LSVRG. 4.…”
Section: D2 Svrcd: Effect Of ρsupporting
confidence: 88%
See 2 more Smart Citations
“…Note that those findings are in accord with Corollary H.3. Similar results were shown in [15] for LSVRG. 4.…”
Section: D2 Svrcd: Effect Of ρsupporting
confidence: 88%
“…Thus, for all j, it does not make sense to increase sampling size beyond point where p i t q t j ≥ 1 n as the convergence speed would not increase significantly 15 .…”
Section: Algorithm 16 Isaega [New Method]mentioning
confidence: 99%
See 1 more Smart Citation
“…in case ii) one may consider the batch gradient ∇f (x k ) if n is small, or a variance-reduced gradient estimator, such as SVRG [28,31] or SAGA [17,46], if n is large. Our general analysis allows for any estimator to be used as long as it satisfies a certain technical assumption (Assumption 2).…”
Section: Summary Of Contributionsmentioning
confidence: 99%
“…Recently, an error compensated method called EC-LSVRG-DIANA which can achieve linear convergence for the strongly convex and smooth case was proposed by Gorbunov et al [2020], but besides the contraction compressor, the unbiased compressor is also needed in the algorithm. In this paper, we study the error compensated methods for loopless SVRG (L-SVRG) [Kovalev et al, 2019], Quartz [Qu et al, 2015], and SDCA [Shalev-Shwartz and Zhang, 2013], where only contraction compressors are needed.…”
Section: Introductionmentioning
confidence: 99%