2016
DOI: 10.1109/tit.2016.2548468
|View full text |Cite
|
Sign up to set email alerts
|

Minimax Rates of Entropy Estimation on Large Alphabets via Best Polynomial Approximation

Abstract: Consider the problem of estimating the Shannon entropy of a distribution over k elements from n independent samples. We show that the minimax mean-square error is within universal multiplicative constant factors of k n log k 2 + log 2 k n if n exceeds a constant factor of k log k ; otherwise there exists no consistent estimator. This refines the recent result of Valiant-Valiant [VV11a] that the minimal sample size for consistent entropy estimation scales according to Θ( k log k ). The apparatus of best polynom… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

6
294
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 166 publications
(300 citation statements)
references
References 58 publications
(48 reference statements)
6
294
0
Order By: Relevance
“…Another estimator proposed by Valiant and Valiant [26] has only been shown to achieve the minimax rate in the restrictive regime of Sln0.2emSnS1.03ln0.2emS. Wu and Yang [27] independently applied the idea of best polynomial approximation to entropy estimation, and obtained its minimax L 2 rate. The minimax lower bound part of Theorem 1 follows from Wu and Yang [27].…”
Section: Introduction and Main Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Another estimator proposed by Valiant and Valiant [26] has only been shown to achieve the minimax rate in the restrictive regime of Sln0.2emSnS1.03ln0.2emS. Wu and Yang [27] independently applied the idea of best polynomial approximation to entropy estimation, and obtained its minimax L 2 rate. The minimax lower bound part of Theorem 1 follows from Wu and Yang [27].…”
Section: Introduction and Main Resultsmentioning
confidence: 99%
“…Wu and Yang [27] independently applied the idea of best polynomial approximation to entropy estimation, and obtained its minimax L 2 rate. The minimax lower bound part of Theorem 1 follows from Wu and Yang [27]. We also remark that, unlike the estimator we propose, the estimator in Wu and Yang [27] relies on knowledge of the support size S , which generally may not be known.…”
Section: Introduction and Main Resultsmentioning
confidence: 99%
“…There exist extensive literature on this subject, and we refer to [22] for a detailed review, as well as the theory and Matlab/Python implementations of entropy and mutual information estimators that achieve the minimax rates in all the regimes of sample size and support size pairs. For the recent growing literature on information measure estimation in the high-dimensional regime, we refer to [23], [24], [25], [22], [26], [27], [28], [29].…”
Section: Problem Formulation and Main Resultsmentioning
confidence: 99%
“…On the other hand, any estimator U E for U can be converted to a (not necessarily good) support size estimator by adding the number of observed symbols. Estimating the support size of an underlying distribution has been studied by both ecologists (1-3) and theoreticians (26)(27)(28); however, to make the problem nontrivial, all statistical models impose a lower bound on the minimum nonzero probability of each symbol, which is assumed to be known to the statistician. We discuss the connections and differences to our results in SI Appendix, section 5.…”
Section: Poissonmentioning
confidence: 99%