2019
DOI: 10.1145/3366700
|View full text |Cite
|
Sign up to set email alerts
|

Fundamental Limits of Approximate Gradient Coding

Abstract: It has been established that when the gradient coding problem is distributed among n servers, the computation load (number of stored data partitions) of each worker is at least s + 1 in order to resists s stragglers [1]. This scheme incurs a large overhead when the number of stragglers s is large. In this paper, we focus on a new framework called approximate gradient coding to mitigate stragglers in distributed learning. We show that, to exactly recover the gradient with high probability, the computation load … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
23
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 37 publications
(23 citation statements)
references
References 17 publications
0
23
0
Order By: Relevance
“…In [22], a distributed training method in the presence of stragglers is presented using the approximate gradient coding of [21], and some reduction in the total training time is reported. In [23], a fundamental trade-off among the computation load, the accuracy of the result, and the number of stragglers is characterized. Furthermore, the authors introduced two schemes to achieve the trade-off.…”
Section: Arxiv:210301589v1 [Csit] 2 Mar 2021mentioning
confidence: 99%
“…In [22], a distributed training method in the presence of stragglers is presented using the approximate gradient coding of [21], and some reduction in the total training time is reported. In [23], a fundamental trade-off among the computation load, the accuracy of the result, and the number of stragglers is characterized. Furthermore, the authors introduced two schemes to achieve the trade-off.…”
Section: Arxiv:210301589v1 [Csit] 2 Mar 2021mentioning
confidence: 99%
“…A second difference between our work and [16] is that [16] does not provide a complete convergence analysis. The works [8], [19], [20] also study schemes similar in flavor to SGC, but these works also do not provide complete convergence analyses.…”
Section: B Relationship To Previous Work On Approximate Gradient Codingmentioning
confidence: 99%
“…We note that BGC is an approximation of the pairwisebalanced schemes we consider and can be seen as a case of SGC when all the data a i have the same norm. In [19] the authors present fundamental bounds on the error as function of the redundancy. In [18], the authors analyze the convergence rate of the fractional repetition scheme presented in [16] and show that under standard assumptions on the loss function, the algorithm maintains the convergence rate of centralized stochastic gradient descent.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The authors of [1] proposed gradient coding, a scheme for exact recovery of the gradient when the objective loss function is additively separable. The exact recovery of the gradient is considered in several prior works, e.g., [1][2][3][4][5][6], while gradient coding for approximate recovery of the gradient is studied in [6][7][8][9][10][11][12][13]. Gradient coding requires the central server to receive the subtasks of a fixed fraction of any of the workers.…”
Section: Introductionmentioning
confidence: 99%