2016
DOI: 10.1109/tac.2016.2529285
|View full text |Cite
|
Sign up to set email alerts
|

Stochastic Gradient-Push for Strongly Convex Functions on Time-Varying Directed Graphs

Abstract: We investigate the convergence rate of the recently proposed subgradient-push method for distributed optimization over time-varying directed graphs. The subgradient-push method can be implemented in a distributed way without requiring knowledge of either the number of agents or the graph sequence; each node is only required to know its out-degree at each time. Our main result is a convergence rate of O ((ln t)/t) for strongly convex functions with Lipschitz gradients even if only stochastic gradient samples ar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
198
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 267 publications
(198 citation statements)
references
References 37 publications
0
198
0
Order By: Relevance
“…In addition, a dual subgradient averaging method is proposed for solving the same problem, which is shown to have better convergence results in terms of network scaling [10]. Some extension has also been made for cases under constraints [11], [12], cases where only noisy gradient is available [13], [14] and directed graphs [15]. A common issue of the abovementioned method is that they require decaying stepsize and the assumption of bounded (sub)gradient to achieve the exact optimum.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition, a dual subgradient averaging method is proposed for solving the same problem, which is shown to have better convergence results in terms of network scaling [10]. Some extension has also been made for cases under constraints [11], [12], cases where only noisy gradient is available [13], [14] and directed graphs [15]. A common issue of the abovementioned method is that they require decaying stepsize and the assumption of bounded (sub)gradient to achieve the exact optimum.…”
Section: Introductionmentioning
confidence: 99%
“…The proposed distributed algorithm allows for using uncoordinated stepsizes for local optimization and, in contrast to most existing literatures, is guaranteed to converge to the exact optimum even with constant stepsizes. Note that our assumption is similar with [7] and [14] except for that we drop the strongly convex assumption but it is quite different from most (sub)gradient-based methods [9], [25], [26] where the (sub)gradients are usually assumed to be bounded, which is quite restrictive in unconstrained optimization problems. It is also important to note that our approach, though, has similar form with the ones proposed in [23], [26]- [28], it differs from them in its nature in that our assumptions (c.f.…”
Section: Introductionmentioning
confidence: 99%
“…We note that more involved gossip protocols have been proposed, we mention for instance broad-cast and push-sum protocols (see [29] and [30]). Although theoretically possible, such an extension of Algorithm 3 is however beyond the scope of this paper.…”
Section: Algorithm 3: Distributed On-line Mle (Domle)mentioning
confidence: 99%
“…Distributed algorithms for solving stochastic problems have been widely studied [6][7][8][9][10][11]. On the other side, distributed algorithms for big-data problems through block communication have started to appear only recently [12][13][14][15][16].…”
Section: Introductionmentioning
confidence: 99%