2015
DOI: 10.3390/e18010013
|View full text |Cite
|
Sign up to set email alerts
|

Improvement of the k-nn Entropy Estimator with Applications in Systems Biology

Abstract: Abstract:In this paper, we investigate efficient estimation of differential entropy for multivariate random variables. We propose bias correction for the nearest neighbor estimator, which yields more accurate results in higher dimensions. In order to demonstrate the accuracy of the improvement, we calculated the corrected estimator for several families of random variables. For multivariate distributions, we considered the case of independent marginals and the dependence structure between the marginal distribut… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
16
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 10 publications
(16 citation statements)
references
References 19 publications
0
16
0
Order By: Relevance
“…A major disadvantage of using KNN entropy estimators to estimate EEG integration and complexity is that even though they are relatively efficient to compute for small multivariate data spaces [ 26 , 29 ], they are computationally intensive to compute for the large data spaces typical of EEG recordings. This is mainly due to the need to perform a SVD for each time point of an EEG epoch as part of the G-KNN computation (Equation (6)) of the joint entropy, H(X) , and conditional entropy, H(X i |X–X i ), and the need to compute the conditional entropy multiple times during the computation of C I (X) (see Equation (2)).…”
Section: Discussionmentioning
confidence: 99%
See 3 more Smart Citations
“…A major disadvantage of using KNN entropy estimators to estimate EEG integration and complexity is that even though they are relatively efficient to compute for small multivariate data spaces [ 26 , 29 ], they are computationally intensive to compute for the large data spaces typical of EEG recordings. This is mainly due to the need to perform a SVD for each time point of an EEG epoch as part of the G-KNN computation (Equation (6)) of the joint entropy, H(X) , and conditional entropy, H(X i |X–X i ), and the need to compute the conditional entropy multiple times during the computation of C I (X) (see Equation (2)).…”
Section: Discussionmentioning
confidence: 99%
“…Fortunately, certain entropy estimators are minimally affected by the nonstationarity of a signal. One such estimator is the nonparametric k-th nearest neighbor (KNN) entropy estimator [ 26 , 27 , 28 , 29 , 30 ], which was used here to compute the entropies composing I(X) and C I (X) . Here, the entropy of a data space is estimated by approximating the multivariate probability density function, f X (X) , at each point, x i , in the space in terms of the ratio of the fraction of other points in a data point’s neighborhood to the volume of its neighborhood [ 29 ]: where N p is the number of data points in the space, Vol i is the volume of the neighborhood, and k(x i ) is the number of points in the neighborhood of x i other than x i .…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…J.Ma and Z.Sun [17] analyzed the copula entropy estimation to get the mutual information estimate. The Gaussian copula in the entropy estimation is also used in [7]. A nonparanormal information estimation is considered in [23].…”
Section: Introductionmentioning
confidence: 99%