2021
DOI: 10.48550/arxiv.2101.11203
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Achieving Linear Speedup with Partial Worker Participation in Non-IID Federated Learning

Haibo Yang,
Minghong Fang,
Jia Liu

Abstract: Federated learning (FL) is a distributed machine learning architecture that leverages a large number of workers to jointly learn a model with decentralized data. FL has received increasing attention in recent years thanks to its data privacy protection, communication efficiency and a linear speedup for convergence in training (i.e., convergence performance increases linearly with respect to the number of workers). However, existing studies on linear speedup for convergence are only limited to the assumptions o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
24
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(24 citation statements)
references
References 16 publications
0
24
0
Order By: Relevance
“…However, FedAvg may not converge if data from different clients is non-i.i.d. (Zhao et al, 2018;Li et al, 2019) and some clients do not regularly participate in the training (Yang et al, 2021), as is often the case in federated learning scenarios. We show similar results for federated graph training.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…However, FedAvg may not converge if data from different clients is non-i.i.d. (Zhao et al, 2018;Li et al, 2019) and some clients do not regularly participate in the training (Yang et al, 2021), as is often the case in federated learning scenarios. We show similar results for federated graph training.…”
Section: Related Workmentioning
confidence: 99%
“…Assumptions 5.1 and 5.2 are standard in the federated learning literature (Yang et al, 2021). We then introduce another assumption to quantify the information loss: Assumption 5.3.…”
Section: Convergence Ratementioning
confidence: 99%
See 1 more Smart Citation
“…Both [19] and [20] consider convergence, but only for strongly convex loss functions. Convergence results for non-convex loss functions with partial device participation have also been presented in [7], [21], [22], but they only consider the case where devices are chosen uniformly at random with or without replacement and do not allow for arbitrary selection probabilities. Finally, [23] considers arbitrary probabilities for each device, but these probabilities are held constant throughout training and are not reflected in the parameter aggregation weights.…”
Section: R Wmentioning
confidence: 99%
“…This is because if (17) holds with an inequality, we can always find an R < R * that satisfies (17) with equality, but the solution (q * , R ) can further reduce the objective function value. Therefore, for the optimal R, (17) always holds, and we can obtain R from (17) and substitute it into the objective of Problem P2. Then, the objective of Problem P2 is…”
Section: B Approximate Optimization Problem For Problem P1mentioning
confidence: 99%