2017
DOI: 10.1137/15m1049695
|View full text |Cite
|
Sign up to set email alerts
|

On the Convergence Rate of Incremental Aggregated Gradient Algorithms

Abstract: Motivated by applications to distributed optimization over networks and large-scale data processing in machine learning, we analyze the deterministic incremental aggregated gradient method for minimizing a finite sum of smooth functions where the sum is strongly convex. This method processes the functions one at a time in a deterministic order and incorporates a memory of previous gradient values to accelerate convergence. Empirically it performs well in practice; however, no theoretical analysis with explicit… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

2
125
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 112 publications
(127 citation statements)
references
References 22 publications
2
125
0
Order By: Relevance
“…In particular, we show that in order to achieve an ǫ-optimal solution, the PIAG algorithm requires O(QK 2 log 2 (QK) log(1/ǫ)) iterations, or equivalently O(QK 2 log(1/ǫ)) iterations, where the tilde is used to hide the logarithmic terms in Q and K. This result improves upon the condition number dependence of the deterministic IAG for smooth problems [12], where the authors proved that to achieve an ǫ-optimal solution, the IAG algorithm requires O(Q 2 K 2 log(1/ǫ)) iterations. We also note that two recent independent papers [9,15] have analyzed the convergence rate of the prox-gradient algorithm (which is a special case of our algorithm with K = 0, i.e., where we have access to a full gradient at each iteration instead of an aggregated gradient) under strong convexity type assumptions and provided linear rate estimates.…”
mentioning
confidence: 75%
See 4 more Smart Citations
“…In particular, we show that in order to achieve an ǫ-optimal solution, the PIAG algorithm requires O(QK 2 log 2 (QK) log(1/ǫ)) iterations, or equivalently O(QK 2 log(1/ǫ)) iterations, where the tilde is used to hide the logarithmic terms in Q and K. This result improves upon the condition number dependence of the deterministic IAG for smooth problems [12], where the authors proved that to achieve an ǫ-optimal solution, the IAG algorithm requires O(Q 2 K 2 log(1/ǫ)) iterations. We also note that two recent independent papers [9,15] have analyzed the convergence rate of the prox-gradient algorithm (which is a special case of our algorithm with K = 0, i.e., where we have access to a full gradient at each iteration instead of an aggregated gradient) under strong convexity type assumptions and provided linear rate estimates.…”
mentioning
confidence: 75%
“…This is in contrast with the recent analysis of the IAG algorithm provided in [12], which used distances of the iterates to the optimal solution as a Lyapunov function and relied on the smoothness of the problem to bound the gradient errors with distances. This approach does not extend to the non-smooth composite case, which motivates a new analysis using function values and the properties of the proximal operator.…”
mentioning
confidence: 94%
See 3 more Smart Citations