2014
DOI: 10.1109/tsp.2013.2291221
|View full text |Cite
|
Sign up to set email alerts
|

Convergence Rates of Distributed Nesterov-Like Gradient Methods on Random Networks

Abstract: Abstract-We consider distributed optimization in random networks where nodes cooperatively minimize the sum of their individual convex costs. Existing literature proposes distributed gradient-like methods that are computationally cheap and resilient to link failures, but have slow convergence rates. In this paper, we propose accelerated distributed gradient methods that 1) are resilient to link failures; 2) computationally cheap; and 3) improve convergence rates over other gradient methods. We model the networ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
46
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
4
3

Relationship

2
5

Authors

Journals

citations
Cited by 44 publications
(46 citation statements)
references
References 38 publications
0
46
0
Order By: Relevance
“…We can modify our methods and relax these prior knowledge requirements such that the methods still provable converge, at rates that are close to the ones presented in this paper; for details, we refer to [10].…”
Section: Algorithm Md-ncmentioning
confidence: 92%
See 3 more Smart Citations
“…We can modify our methods and relax these prior knowledge requirements such that the methods still provable converge, at rates that are close to the ones presented in this paper; for details, we refer to [10].…”
Section: Algorithm Md-ncmentioning
confidence: 92%
“…In subsequent results, ξ denotes an arbitrarily small positive number. A proof of Theorem 1, as well as explicit constants in the established rates, can be found in [10]. Theorem 1 indicates that the convergence rates do not depend on the underlying random network statistics.…”
Section: Convergence Analysismentioning
confidence: 93%
See 2 more Smart Citations
“…We consider standard distributed stochastic gradient methods where at each time step, each node makes a weighted average of its own and its neighbors' solution estimates, and performs a step in the negative direction of its noisy local gradient. The underlying network is allowed to be randomly varying, similarly to, e.g., the models in [4]- [6]. More specifically, the network is modeled through a sequence of independent identically distributed (i.i.d.)…”
Section: Introductionmentioning
confidence: 99%