ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9053448
|View full text |Cite
|
Sign up to set email alerts
|

Cooperative Learning VIA Federated Distillation OVER Fading Channels

Abstract: Cooperative training methods for distributed machine learning are typically based on the exchange of local gradients or local model parameters. The latter approach is known as Federated Learning (FL). An alternative solution with reduced communication overhead, referred to as Federated Distillation (FD), was recently proposed that exchanges only averaged model outputs. While prior work studied implementations of FL over wireless fading channels, here we propose wireless protocols for FD and for an enhanced ver… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
26
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 32 publications
(27 citation statements)
references
References 19 publications
1
26
0
Order By: Relevance
“…While FD performs well when the mobile devicegenerated data is identically and independently distributed, FD exhibits lower performance than the FL benchmark with model parameter exchange in non-IID data distributions. This was experimentally verified in [6], [7], [8], [9] and the experiments presented in Section 4. To fill this gap, we design an FL framework with model output exchange achieving similar or higher performance than previously proposed approaches even when subjected to non-IID data distributions.…”
Section: Related Work and Paper Organizationsupporting
confidence: 58%
See 2 more Smart Citations
“…While FD performs well when the mobile devicegenerated data is identically and independently distributed, FD exhibits lower performance than the FL benchmark with model parameter exchange in non-IID data distributions. This was experimentally verified in [6], [7], [8], [9] and the experiments presented in Section 4. To fill this gap, we design an FL framework with model output exchange achieving similar or higher performance than previously proposed approaches even when subjected to non-IID data distributions.…”
Section: Related Work and Paper Organizationsupporting
confidence: 58%
“…Federated learning with model output exchange over mobile device-generated dataset. FD is proposed in [6], [7], [8], [9] as an FL framework with model output exchange that trains ML models considering mobile devicegenerated dataset. Unlike CD and PATE that trains distributed ML models using a shared dataset, each mobile device trains each ML model using a local dataset, enabling ML model training with mobile device-generated data.…”
Section: Related Work and Paper Organizationmentioning
confidence: 99%
See 1 more Smart Citation
“…4(b) (where 0 < λ < 1) into Eq. (7). The dependence of the training loss is insignificant at any K, even when the transmission rate was reduced from 6 Mbps to 3 Mbps.…”
Section: B Effects Of Network Density On C-sgdmentioning
confidence: 90%
“…Some researchers have investigated methods for quantization and sparsification of data that must be communicated to reduce the communication load. In particular, compressed sensing [6], [29] and digital data coding [7], [8] have been explored for data reduction when communicating with a central server. These approaches have been extended to be applicable to wireless systems.…”
Section: Related Work a Centralized Settingmentioning
confidence: 99%