2012
DOI: 10.1007/978-3-642-32717-9_21
|View full text |Cite
|
Sign up to set email alerts
|

Information Theoretic Clustering Using Minimum Spanning Trees

Abstract: Abstract. In this work we propose a new information-theoretic clustering algorithm that infers cluster memberships by direct optimization of a non-parametric mutual information estimate between data distribution and cluster assignment. Although the optimization objective has a solid theoretical foundation it is hard to optimize. We propose an approximate optimization formulation that leads to an efficient algorithm with low runtime complexity. The algorithm has a single free parameter, the number of clusters t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
20
0

Year Published

2014
2014
2021
2021

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 36 publications
(20 citation statements)
references
References 14 publications
0
20
0
Order By: Relevance
“…In fact I S = 0 if and only if the two studied random variable are strictly independent. Mutual information is then a natural measure which can be used to extend the similarity measure to make it sensitive to non-linear dependencies, and has been successfully used in some applications [45][46][47]. Recently we have used it in the creation of dependency networks on financial markets [20].…”
Section: Introductionmentioning
confidence: 99%
“…In fact I S = 0 if and only if the two studied random variable are strictly independent. Mutual information is then a natural measure which can be used to extend the similarity measure to make it sensitive to non-linear dependencies, and has been successfully used in some applications [45][46][47]. Recently we have used it in the creation of dependency networks on financial markets [20].…”
Section: Introductionmentioning
confidence: 99%
“…x − x , in which · can be any norm. 2 and ∞ are commonly used. c dx is the volume of corresponding unit norm ball.…”
Section: Kl Entropy Estimatormentioning
confidence: 99%
“…Information theoretic quantities, such as Shannon entropy and mutual information, have a broad range of applications in statistics and machine learning, such as clustering [2,3], feature selection [4,5], anomaly detection [6], test of normality [7], etc. These quantities are determined by the distributions of random variables, which are usually unknown in real applications.…”
Section: Introductionmentioning
confidence: 99%
“…To address this problem, we propose to calculate the global and local density of the point. Some MST-based algorithms are combined with other methods, such as information theory [16], k-means [17], and multivariate Gaussians [18].…”
Section: Related Workmentioning
confidence: 99%