2015
DOI: 10.48550/arxiv.1507.02000
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

An optimal randomized incremental gradient method

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

3
58
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 31 publications
(61 citation statements)
references
References 17 publications
3
58
0
Order By: Relevance
“…To the best of our knowledge, ANITA is the first accelerated algorithm which can exactly achieve this optimal result O n + nL for general convex finite-sum problems. In the strongly convex setting, we also show that ANITA can achieve the optimal convergence result O n + nL µ log 1 matching the lower bound Ω n + nL µ log 1 provided by Lan and Zhou (2015). Moreover, ANITA enjoys a simpler loopless algorithmic structure unlike previous accelerated algorithms such as Katyusha (Allen-Zhu, 2017) and Varag (Lan et al, 2019) where they use an inconvenient double-loop structure.…”
supporting
confidence: 59%
See 2 more Smart Citations
“…To the best of our knowledge, ANITA is the first accelerated algorithm which can exactly achieve this optimal result O n + nL for general convex finite-sum problems. In the strongly convex setting, we also show that ANITA can achieve the optimal convergence result O n + nL µ log 1 matching the lower bound Ω n + nL µ log 1 provided by Lan and Zhou (2015). Moreover, ANITA enjoys a simpler loopless algorithmic structure unlike previous accelerated algorithms such as Katyusha (Allen-Zhu, 2017) and Varag (Lan et al, 2019) where they use an inconvenient double-loop structure.…”
supporting
confidence: 59%
“…• Finally, for strongly convex finite-sum problems (i.e., Assumption 2 holds), we also prove that ANITA achieves the optimal convergence result O n+ nL µ log 1 matching the lower bound Ω n+ nL µ log 1 provided by Lan and Zhou (2015) (see Table 1). b ANITA can achieve this optimal result for a very wide range of , i.e., ∈ (0, 2 for more details).…”
Section: Our Contributionsmentioning
confidence: 63%
See 1 more Smart Citation
“…Acceleration of gradient-type methods is widely-studied for standard optimization problems (Lan and Zhou, 2015;Lin et al, 2015;Allen-Zhu, 2017;Lan et al, 2019;Li, 2021a). Deep learning practitioners typically rely on Adam (Kingma and Ba, 2014), or one of its many variants, which besides other tricks also adopts momentum.…”
Section: Methods With Accelerationmentioning
confidence: 99%
“…◇ Momentum. A very successful and popular technique for enhancing both optimization and generalization is momentum/acceleration (Polyak, 1964;Nesterov, 1983;Lan & Zhou, 2015;Allen-Zhu, 2017;Lan et al, 2019;Li, 2021a). For instance, momentum is a key building block behind the widely-used Adam method (Kingma & Ba, 2014).…”
Section: Our Contributionsmentioning
confidence: 99%