IEEE INFOCOM 2019 - IEEE Conference on Computer Communications 2019
DOI: 10.1109/infocom.2019.8737489
|View full text |Cite
|
Sign up to set email alerts
|

Compressed Distributed Gradient Descent: Communication-Efficient Consensus over Networks

Abstract: Network consensus optimization has received increasing attention in recent years and has found important applications in many scientific and engineering fields. To solve network consensus optimization problems, one of the most well-known approaches is the distributed gradient descent method (DGD). However, in networks with slow communication rates, DGD's performance is unsatisfactory for solving high-dimensional network consensus problems due to the communication bottleneck. This motivates us to design a commu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 25 publications
(22 citation statements)
references
References 28 publications
0
22
0
Order By: Relevance
“…The resulting global coefficient matrix W ∈ N ×N with entries w kl is right-stochastic and also known as consensus matrix, which has been proven to reach a consistent estimation for all nodes over network [19]. The diffusion GN method aims to enhance the interaction among neighboring nodes by aggregating the real time estimates from neighborhood as (13) and integrating them into local update step as (14), such that the network-wide sensing data is fused by individual node as long as the network is connected.…”
Section: Diffusion Gauss-newton Methods For Node Localizationmentioning
confidence: 99%
“…The resulting global coefficient matrix W ∈ N ×N with entries w kl is right-stochastic and also known as consensus matrix, which has been proven to reach a consistent estimation for all nodes over network [19]. The diffusion GN method aims to enhance the interaction among neighboring nodes by aggregating the real time estimates from neighborhood as (13) and integrating them into local update step as (14), such that the network-wide sensing data is fused by individual node as long as the network is connected.…”
Section: Diffusion Gauss-newton Methods For Node Localizationmentioning
confidence: 99%
“…There are four terms in the convergence error of SDM-DSGD in Eq. 7: (I) is the common convergence error that goes to zero as ) and step-size \W increase; (II) is the approximation error between the Lyapunov function + W (x; D) and f (x; D), which decreases with W. These two terms are similar to those in the convergence of DGD-based algorithms [11,28]; (III) and (IV) are the error terms introduced by the compression, random sampling, as well as the Gaussian masking noises. The following simplied convergence rate result follows immediately from Lemma 2.…”
Section: (C) Under Assumption 1mentioning
confidence: 98%
“…R 1. Algorithm 1 is motivated by and bears some similarity with the DGD-type communication-ecient distributed learning in the literature [7,11]. In these existing work, rather than exchanging the states directly, the compressed dierentials between two successive iterations of the variables are communicated to reduce the communication load.…”
Section: The Sdm-dsgd Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…The use cases of such algorithms are internet-of-things, peer-topeer (P2P) network sharing, vehicular communication, sensor networks, edge devices, etc [17], [18]. There exist several works applying decentralized optimization methods for neural networks without a master node [19]- [23]. For instance, the work [19] uses amplified-differential compression for computing the gradient in a synchronous manner and enjoys low computational complexity.…”
Section: Introductionmentioning
confidence: 99%