2021
DOI: 10.48550/arxiv.2111.14655
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

FedHM: Efficient Federated Learning for Heterogeneous Models via Low-rank Factorization

Abstract: The underlying assumption of recent federated learning (FL) paradigms is that local models usually share the same network architecture as the global model, which becomes impractical for mobile and IoT devices with different setups of hardware and infrastructure. A scalable federated learning framework should address heterogeneous clients equipped with different computation and communication capabilities. To this end, this paper proposes FEDHM, a novel federated model compression framework that distributes the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(11 citation statements)
references
References 15 publications
0
10
0
Order By: Relevance
“…The major difference to its use for inference is that here, the low-rank NN is updated during training, and the low-rank updates are applied to the full model on the server. Yao et al [101] present FedHM, where they create low-complexity submodels on the server by doing a low-rank factorization of the full model. Layer parameters with dimensions 𝑚 × 𝑛 are decomposed into two matrices with dimensions 𝑚 × 𝑟 and 𝑟 × 𝑛.…”
Section: Nn Architecture Heterogeneity Based On Fedavgmentioning
confidence: 99%
See 1 more Smart Citation
“…The major difference to its use for inference is that here, the low-rank NN is updated during training, and the low-rank updates are applied to the full model on the server. Yao et al [101] present FedHM, where they create low-complexity submodels on the server by doing a low-rank factorization of the full model. Layer parameters with dimensions 𝑚 × 𝑛 are decomposed into two matrices with dimensions 𝑚 × 𝑟 and 𝑟 × 𝑛.…”
Section: Nn Architecture Heterogeneity Based On Fedavgmentioning
confidence: 99%
“…The attributes scale and granularity are often neglected, are hidden behind the technique, and lack discussion in the papers. The reported scale in the resources supported by the techniques ranges from 4× − 25× [12,41,52,61,71,77,79,80,85,101] up to 100× − 250× [25,87], yet it remains unclear whether training at such high scales is still effective. Hence, while all approaches show the effectiveness of their solution in certain scenarios, it often remains unclear whether devices with low resources or stale devices can make a meaningful contribution that advances the global model.…”
Section: Open Problems and Future Directionsmentioning
confidence: 99%
“…Such edge devices are typically resource-constrained, e.g., the computing, communication, and memory capacities are limited. Several research efforts have been conducted to enhance the computation and communication efficiency of cross-device FL via model updates sparsification, quantization, and low-rank factorization [25,51,56,21]. Training deep neural networks requires high memory consumption [47].…”
Section: Related Workmentioning
confidence: 99%
“…As such, lately, there has been a line of work focusing on this very problem, where the discrepancy between the dynamics of different clients affects the convergence rate or fairness of the deployed system. Specifically, such solutions draw from efficient ML and attempt to dynamically alter the footprint of local models my means of structured (PruneFL [37]), unstructured (Adaptive Federated Dropout [13]) or importance-based pruning (FjORD [33]), quantisation (AQFL [2]), low-rank factorisation (FedHM [76]), sparsity-inducing training (ZeroFL [63]) or distillation (GKT [31]). However, each approach has limitations, either because they involve extra training overhead [31] and residence of multiple DNN copies in memory [2], or because they require specialised hardware for performance gains ( [63,13]).…”
Section: Related Workmentioning
confidence: 99%
“…However, each approach has limitations, either because they involve extra training overhead [31] and residence of multiple DNN copies in memory [2], or because they require specialised hardware for performance gains ( [63,13]). Last, some of the architectural changes proposed may not offer the degrees of freedom that NAS exposes [33,76].…”
Section: Related Workmentioning
confidence: 99%