2019 57th Annual Allerton Conference on Communication, Control, and Computing (Allerton) 2019
DOI: 10.1109/allerton.2019.8919853
|View full text |Cite
|
Sign up to set email alerts
|

Nested Distributed Gradient Methods with Stochastic Computation Errors

Abstract: In this work, we consider the problem of a network of agents collectively minimizing a sum of convex functions. The agents in our setting can only access their local objective functions and exchange information with their immediate neighbors. Motivated by applications where computation is imperfect, including, but not limited to, empirical risk minimization (ERM) and online learning, we assume that only noisy estimates of the local gradients are available. To tackle this problem, we adapt a class of Nested Dis… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

1
2
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2

Relationship

2
0

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 48 publications
1
2
0
Order By: Relevance
“…Our method is based on a class of flexible algorithms (NEAR-DGD) [57] that permit the trade-off of computation and communication to best accommodate the application setting. In this work, we generalize our previous results analyzing NEAR-DGD in the presence of either deterministically quantized communication [58] or stochastic gradient errors [59], and unify them under a common, fully stochastic framework. We provide theoretical results to demonstrate that S-NEAR-DGD converges to a neighborhood of the optimal solution with geometric rate, and that if an error-correction mechanism is incorporated to consensus, then the total communication error induced by inexact communication is independent of the number of consensus rounds peformed by our algorithm.…”
Section: B Contributionssupporting
confidence: 56%
See 1 more Smart Citation
“…Our method is based on a class of flexible algorithms (NEAR-DGD) [57] that permit the trade-off of computation and communication to best accommodate the application setting. In this work, we generalize our previous results analyzing NEAR-DGD in the presence of either deterministically quantized communication [58] or stochastic gradient errors [59], and unify them under a common, fully stochastic framework. We provide theoretical results to demonstrate that S-NEAR-DGD converges to a neighborhood of the optimal solution with geometric rate, and that if an error-correction mechanism is incorporated to consensus, then the total communication error induced by inexact communication is independent of the number of consensus rounds peformed by our algorithm.…”
Section: B Contributionssupporting
confidence: 56%
“…We considered quantized communication using deterministic (D) algorithms (e.g. rounding to the nearest integer with no uncertainty) in [58], while a variant of NEAR-DGD that utilizes stochastic gradient approximations only was presented in [59]. This work unifies and generalizes these methods.…”
Section: Methods Communication Computationmentioning
confidence: 99%
“…1. distributed first-order primal algorithms [15][16][17][18][19][20][21][22][23][24]: arXiv:2006.01665v1 [math.OC] 31 May 2020 methods that use only gradient information and operate in primal space (i.e., directly on problem (1.3)); 2. nested [1,[25][26][27][28]: methods that decompose the communication and computation steps and perform them sequentially;…”
Section: Introductionmentioning
confidence: 99%