2014
DOI: 10.14209/jcis.2014.8
|View full text |Cite
|
Sign up to set email alerts
|

A pragmatic entropy and differential entropy estimator for small datasets

Abstract: Abstract-A pragmatic approach for entropy estimation is presented, first for discrete variables, then in the form of an extension for handling continuous and/or multivariate ones. It is based on coincidence detection, and its application leads to algorithms with three main attractive features: they are easy to use, can be employed without any a priori knowledge concerning source distribution (not even the alphabet cardinality K of discrete sources) and can provide useful estimates even when the number of sampl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(12 citation statements)
references
References 9 publications
0
12
0
Order By: Relevance
“…continuous L−dimensional multivariate data coupled to categorical data) they should be separately preprocessed for proper definition of multivariate coincidence event. As proposed in [19], in the Section entitled Setting and testing a coincidence neighbourhood, if the standard deviations in each dimension of the continuous data are normalized, one may take advantage of an approximative relationship between a first guess for the effective cardinality and the hyper-volume ∆. For instance, given a first rough guess for effective cardinality of about C (2) = 1000, and L = 3, then one may set ∆ ≈…”
Section: Discussionmentioning
confidence: 99%
“…continuous L−dimensional multivariate data coupled to categorical data) they should be separately preprocessed for proper definition of multivariate coincidence event. As proposed in [19], in the Section entitled Setting and testing a coincidence neighbourhood, if the standard deviations in each dimension of the continuous data are normalized, one may take advantage of an approximative relationship between a first guess for the effective cardinality and the hyper-volume ∆. For instance, given a first rough guess for effective cardinality of about C (2) = 1000, and L = 3, then one may set ∆ ≈…”
Section: Discussionmentioning
confidence: 99%
“…Besides, [26] presents entropy as an effective cardinality in logarithmic scale. Likewise, DE can be regarded as an effective volume (in logarithmic scale) [20], [25].…”
Section: Intrinsic Dimension and Differential Entropy In A Nutshellmentioning
confidence: 99%
“…On the other hand, in [19], the problem of DE estimation in high-dimensional spaces was tackled through a simple but data-efficient approach, referred to as the Coincidence Method (CM), originally applied in Physics. In [20] this method was extended to differential entropy estimation in the pattern recognition context, which clearly shows that the correlation dimension in [14] uses the same empirical coincidence ratio as the entropy estimation method proposed in [19].…”
Section: Introductionmentioning
confidence: 99%
“…Since Φ(−z) ≤ 1 √ 2πz e − z 2 2 for z > 1, taking into account inequalities (26), (28) and (35) we get, for k = k n ∝ n α , x ∈ B 1,n and all n large enough,…”
Section: Proof Of Theoremmentioning
confidence: 99%
“…Note that the probability distribution discretization techniques for a random variable having a density (w.r.t. the Lebesgue measure) and evaluation of the Shannon entropy for thus arising random variables do not lead to the differential entropy as the mash of the discretization tends to zero (see, e.g., Theorem 8.3.1 in [13] and [28]). More generally, when a measure σ is fixed on a measure space (S, B), one can define the notion of the entropy of a probability measure ν given on the same space and absolutely continuous w.r.t.…”
mentioning
confidence: 99%