2016 IEEE Global Communications Conference (GLOBECOM) 2016
DOI: 10.1109/glocom.2016.7841903
|View full text |Cite
|
Sign up to set email alerts
|

Information Theoretic Limits of Data Shuffling for Distributed Learning

Abstract: Abstract-Data shuffling is one of the fundamental building blocks for distributed learning algorithms, that increases the statistical gain for each step of the learning process. In each iteration, different shuffled data points are assigned by a central node to a distributed set of workers to perform local computations, which leads to communication bottlenecks. The focus of this paper is on formalizing and understanding the fundamental information-theoretic tradeoff between storage (per worker) and the worst-c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
34
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 38 publications
(34 citation statements)
references
References 4 publications
0
34
0
Order By: Relevance
“…In [75], the authors study the information-theoretic limits of the coded shuffling problem. More specifically, the authors completely characterize the fundamental limits for the case of 2 workers and the case of 3 workers.…”
Section: B Data Shuffling and Communication Overheadsmentioning
confidence: 99%
“…In [75], the authors study the information-theoretic limits of the coded shuffling problem. More specifically, the authors completely characterize the fundamental limits for the case of 2 workers and the case of 3 workers.…”
Section: B Data Shuffling and Communication Overheadsmentioning
confidence: 99%
“…In other words, τ * is the time for which there are exactly r inner products -on average -aggregated at the master node, when the workers are loaded according to the loading obtained in (13). Using (13), (14) and (16), we find that…”
Section: B Solving the Alternate Formulationmentioning
confidence: 99%
“…where K|Q and each node is supposed to compute Q K functions. In fact, in this setup, the K in definitions (21) and (22) will be replaced by Q. Achievability proofs can be shown by executing Q K times the distributed computing schemes as explained in Section IV and V. The converse can be derived by adjusting the definitions in (119), (121), (194), and following the same steps in Appendix B.…”
Section: Definition 4 (Fundamental Scc Regionmentioning
confidence: 99%