2020
DOI: 10.48550/arxiv.2010.15088
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Finite-Time Analysis of Decentralized Stochastic Approximation with Applications in Multi-Agent and Multi-Task Learning

Abstract: Stochastic approximation, a data-driven approach for finding the fixed point of an unknown operator, provides a unified framework for treating many problems in stochastic optimization and reinforcement learning. Motivated by a growing interest in multi-agent and multi-task learning, we consider in this paper a decentralized variant of stochastic approximation. A network of agents, each with their own unknown operator and data observations, cooperatively find the fixed point of the aggregate operator. The agent… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

1
9
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(10 citation statements)
references
References 0 publications
1
9
0
Order By: Relevance
“…The first three papers here deal with the distributed TD(0) method. These show that a result similar to the one in Zeng et al [2020b] holds for this method under constant stepsizes. In contrast, when α n is of the 1/n type, it is proven that E…”
Section: Related Worksupporting
confidence: 76%
See 2 more Smart Citations
“…The first three papers here deal with the distributed TD(0) method. These show that a result similar to the one in Zeng et al [2020b] holds for this method under constant stepsizes. In contrast, when α n is of the 1/n type, it is proven that E…”
Section: Related Worksupporting
confidence: 76%
“…Finite-time literature, in contrast, majorly talks about expectation bounds. Assuming there exists a unique x * that solves m i=1 h i (x) = 0, these results describe the rate at which E x n − x * decays with n. Notable contributions for DSA here are Zeng et al [2020b], Wai [2020]. Compared to ours, these look at a slightly different setup: the measurement noise at each node has a Markov component in place of a martingale difference term.…”
Section: Related Workmentioning
confidence: 77%
See 1 more Smart Citation
“…This can be viewed as an application of ideas from distributed stochastic approximation [34][35][36][37][38][39]. Finite-time performance guarantees for distributed RL have also been provided in works, most notably in [3,[40][41][42][43][44].…”
Section: Related Workmentioning
confidence: 99%
“…The assumption that is the central concern of this paper and is made in all the existing finite-time analyses for distributed RL algorithms is that the consensus interaction is characterized by doubly stochastic matrices [3,[40][41][42][43][44] at every time step, or at least in expectation, i.e., W 1 = 1 and 1 ⊤ E(W ) = 1 ⊤ [38]. Intuitively, doubly stochastic matrices imply symmetry in the communication graph, which almost always requires bidirectional communication graphs.…”
Section: Related Workmentioning
confidence: 99%