2022
DOI: 10.48550/arxiv.2208.02507
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ZeroFL: Efficient On-Device Training for Federated Learning with Local Sparsity

Abstract: When the available hardware cannot meet the memory and compute requirements to efficiently train high performing machine learning models, a compromise in either the training quality or the model complexity is needed. In Federated Learning (FL), nodes are orders of magnitude more constrained than traditional servergrade hardware and are often battery powered, severely limiting the sophistication of models that can be trained under this paradigm. While most research has focused on designing better aggregation st… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 25 publications
0
1
0
Order By: Relevance
“…Moreover, in Li et al (2020a), each client trains a personalized mask to maximize the performance only on the local data. A few recent works Bibikar et al (2022); Huang et al (2022); Qiu et al (2022); Li et al (2020a) also attempted to leverage sparse training within the FL setting as well. In particular Li et al (2020a) implemented randomly initialized sparse mask, FedDST Bibikar et al (2022) built on the idea of RigL Evci et al (2020) and mostly focussed on magnitude pruning on the server-side resulting in similar constraints and Ohib et al (2023) uses sparse gradients to efficiently train in a federated learning setting.…”
Section: Efficiency In Federated Learningmentioning
confidence: 99%
“…Moreover, in Li et al (2020a), each client trains a personalized mask to maximize the performance only on the local data. A few recent works Bibikar et al (2022); Huang et al (2022); Qiu et al (2022); Li et al (2020a) also attempted to leverage sparse training within the FL setting as well. In particular Li et al (2020a) implemented randomly initialized sparse mask, FedDST Bibikar et al (2022) built on the idea of RigL Evci et al (2020) and mostly focussed on magnitude pruning on the server-side resulting in similar constraints and Ohib et al (2023) uses sparse gradients to efficiently train in a federated learning setting.…”
Section: Efficiency In Federated Learningmentioning
confidence: 99%
“…Tensor pruning involves setting entire blocks of the model to zero, which can be useful when the model has a structured block-wise pattern. Structured pruning can preserve the underlying structure of the model and can result in higher compression rates than random pruning [195]. Still, it may require more complex implementation and may introduce more errors in the model or gradient values.…”
mentioning
confidence: 99%