2018
DOI: 10.48550/arxiv.1806.00582
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Federated Learning with Non-IID Data

Abstract: Federated learning enables resource-constrained edge compute devices, such as mobile phones and IoT devices, to learn a shared model for prediction, while keeping the training data local. This decentralized approach to train models provides privacy, security, regulatory and economic benefits. In this work, we focus on the statistical challenge of federated learning when local data is non-IID. We first show that the accuracy of federated learning reduces significantly, by up to ~55% for neural networks trained … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

11
824
1
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 642 publications
(971 citation statements)
references
References 18 publications
11
824
1
1
Order By: Relevance
“…The first FL framework, Federated Averaging (Fe-dAvg) (McMahan et al 2017) glb for the next training round. Although FedAvg achieves superior performance on homogeneous clients with IID data, its performance degrades when data distribution among client devices is non-IID (Zhao et al 2018;Wang et al 2020a). While multiple FL frameworks were proposed to mitigate the impact of statistical diversity by either controlling the divergence between the local model and global model (Li et al 2020;Karimireddy et al 2019), or by training a personalized model for each client (Smith et al 2017;Deng, Kamani, and Mahdavi 2020;Zhang et al 2020), none of these studies has considered system performance objectives in their design.…”
Section: Federated Learningmentioning
confidence: 99%
“…The first FL framework, Federated Averaging (Fe-dAvg) (McMahan et al 2017) glb for the next training round. Although FedAvg achieves superior performance on homogeneous clients with IID data, its performance degrades when data distribution among client devices is non-IID (Zhao et al 2018;Wang et al 2020a). While multiple FL frameworks were proposed to mitigate the impact of statistical diversity by either controlling the divergence between the local model and global model (Li et al 2020;Karimireddy et al 2019), or by training a personalized model for each client (Smith et al 2017;Deng, Kamani, and Mahdavi 2020;Zhang et al 2020), none of these studies has considered system performance objectives in their design.…”
Section: Federated Learningmentioning
confidence: 99%
“…However, FedAvg may not converge if data from different clients is non-i.i.d. (Zhao et al, 2018;Li et al, 2019) and some clients do not regularly participate in the training (Yang et al, 2021), as is often the case in federated learning scenarios. We show similar results for federated graph training.…”
Section: Related Workmentioning
confidence: 99%
“…This data, however, is often privacy-sensitive: for example, users in a social network may not want to reveal the websites they have visited. In non-graphical settings, distributed learning has recently shown promise for preserving user privacy while training accurate models, e.g., federated learning algorithms have become increasingly popular (Yang et al, 2021;Zhao et al, 2018). Some papers have begun to apply federated algo-1 Code available at https://github.com/yh-yao/Federated-GCN rithms to training GCNs (He et al, 2021;Wang et al, 2020).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In ideal cases, Independent and Identically Distributed (IID) data are desired, both in terms of features as well as classes. However, in real scenarios, non-IID data varying from moderate to strongly skewed distribution is inevitable [5,6]. The global class imbalance problem may prevail and adversely affect the performance of the global model [7].…”
Section: Introductionmentioning
confidence: 99%