2019 IEEE 21st International Conference on High Performance Computing and Communications; IEEE 17th International Conference On 2019
DOI: 10.1109/hpcc/smartcity/dss.2019.00046
|View full text |Cite
|
Sign up to set email alerts
|

Entropy-Based Gradient Compression for Distributed Deep Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(9 citation statements)
references
References 14 publications
0
7
0
Order By: Relevance
“…However, network communication is the major problem of DDL. Several methods [14][15][16][29][30][31][32] can be used to reduce the amount of network traffic, but this comes at a cost in terms of accuracy.…”
Section: Discussionmentioning
confidence: 99%
“…However, network communication is the major problem of DDL. Several methods [14][15][16][29][30][31][32] can be used to reduce the amount of network traffic, but this comes at a cost in terms of accuracy.…”
Section: Discussionmentioning
confidence: 99%
“…Using the obtained entropy information and QuickSelect algorithm, the threshold is calculated and only those gradients with absolute value above the threshold are transmitted in that communication round. The results in [45] showed that up to 1000 times gradient compression is achievable while keeping the accuracy of the model nearly unchanged. Fast FL was proposed by Nori et al [46] which attempts to jointly consider the local weight updates and gradient compression tradeoff in FL.…”
Section: B Gradient Compressionmentioning
confidence: 99%
“…Abrahamyan et al [44] designed an autoencoder with a lightweight architecture which captures the common patterns in the gradients of the different distributed clients and achieved a 8095 times compression which is 8 times more than DGC. Entropy based gradient compression scheme was proposed by Kuang et al [45] which consisted of an entropy based threshold selection method and a learning rate correction algorithm. Entropy is a well known metric from information theory which here measures the uncertainty or disorder of the gradients.…”
Section: B Gradient Compressionmentioning
confidence: 99%
“…Abrahamyan et al 44 designed an autoencoder with a lightweight architecture which captures the common patterns in the gradients of the different distributed clients and achieved a 8095 times compression which is 8 times more than DGC. Entropy based gradient compression scheme was proposed by Kuang et al 45 which consisted of an entropy based threshold selection method and a learning rate correction algorithm. Entropy is a well-known metric from information theory which here measures the uncertainty or disorder of the gradients.…”
Section: Gradient Compressionmentioning
confidence: 99%