Proceedings of the Forty-Eighth Annual ACM Symposium on Theory of Computing 2016
DOI: 10.1145/2897518.2897582
|View full text |Cite
|
Sign up to set email alerts
|

Communication lower bounds for statistical estimation problems via a distributed data processing inequality

Abstract: We study the tradeoff between the statistical error and communication cost of distributed statistical estimation problems in high dimensions. In the distributed sparse Gaussian mean estimation problem, each of the m machines receives n data points from a d-dimensional Gaussian distribution with unknown mean θ which is promised to be k-sparse. The machines communicate by message passing and aim to estimate the mean θ. We provide a tight (up to logarithmic factors) tradeoff between the estimation error and the n… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

2
131
1

Year Published

2017
2017
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 112 publications
(134 citation statements)
references
References 10 publications
2
131
1
Order By: Relevance
“…More recently, Duchi and Rogers [16] showed how to combine the above analogue of Assouad's method with techniques from information complexity [8,22] to prove lower bounds for estimation problems that apply to a restricted class of fully interactive locally private protocols. A corollary of their lower bounds is that several known noninteractive algorithms are optimal minimax estimators within the class they consider.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…More recently, Duchi and Rogers [16] showed how to combine the above analogue of Assouad's method with techniques from information complexity [8,22] to prove lower bounds for estimation problems that apply to a restricted class of fully interactive locally private protocols. A corollary of their lower bounds is that several known noninteractive algorithms are optimal minimax estimators within the class they consider.…”
Section: Related Workmentioning
confidence: 99%
“…Let h 2 denote the square of the Hellinger distance, h 2 (f, g) = 1 − X f (x)g(x)dx. We begin with Lemma 5.4, originally proven as Lemma 2 in Braverman et al [8].…”
mentioning
confidence: 99%
“…Xu and Raginsky [16] provided lower bounds on the risk in a distributed Bayesian estimation setting with noisy channels between the data collection terminals and the estimation entity. Braverman et al [17] provided lower bounds for some high dimensional distributed estimation problems, again when the samples of all terminals are from the same distribution, e.g. for distributed estimation of the multivariate Guassian mean when it is known to be sparse.…”
Section: Related Workmentioning
confidence: 99%
“…The distributed mean estimation problem was recently studied in a statistical framework where it is assumed that the vectors X i are independent and identicaly distributed samples from some specific underlying distribution. In such a setup, the goal is to estimate the true mean of the underlying distribution [14,13,2,1]. These works formulate lower and upper bounds on the communication cost needed to achieve the minimax optimal estimation error.…”
Section: Background and Contributionsmentioning
confidence: 99%