2022
DOI: 10.48550/arxiv.2203.05468
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

CoCo-FL: Communication- and Computation-Aware Federated Learning via Partial NN Freezing and Quantization

Abstract: Devices participating in federated learning (FL) typically have heterogeneous communication and computation resources. However, all devices need to finish training by the same deadline dictated by the server when applying synchronous FL, as we consider in this paper. Reducing the complexity of the trained neural network (NN) at constrained devices, i.e., by dropping neurons/filters, is insufficient as it tightly couples reductions in communication and computation requirements, wasting resources. Quantization h… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 15 publications
0
3
0
Order By: Relevance
“…In ZeroFL [79], dropout masks in combination with sparse convolutions are used to lower the computational complexity in training (FLOPS) and reduce the communication volume, although special hardware and software support is required to enable real-world gains. Lastly, in CoCoFL [77], a technique is presented that does not use subsets of an NN for training. Instead, only for some layers per round gradients get calculated while the remainder of the layers are frozen.…”
Section: Othersmentioning
confidence: 99%
See 2 more Smart Citations
“…In ZeroFL [79], dropout masks in combination with sparse convolutions are used to lower the computational complexity in training (FLOPS) and reduce the communication volume, although special hardware and software support is required to enable real-world gains. Lastly, in CoCoFL [77], a technique is presented that does not use subsets of an NN for training. Instead, only for some layers per round gradients get calculated while the remainder of the layers are frozen.…”
Section: Othersmentioning
confidence: 99%
“…The attributes scale and granularity are often neglected, are hidden behind the technique, and lack discussion in the papers. The reported scale in the resources supported by the techniques ranges from 4× − 25× [12,41,52,61,71,77,79,80,85,101] up to 100× − 250× [25,87], yet it remains unclear whether training at such high scales is still effective. Hence, while all approaches show the effectiveness of their solution in certain scenarios, it often remains unclear whether devices with low resources or stale devices can make a meaningful contribution that advances the global model.…”
Section: Open Problems and Future Directionsmentioning
confidence: 99%
See 1 more Smart Citation