2001
DOI: 10.1007/3-540-48219-9_31
|View full text |Cite
|
Sign up to set email alerts
|

Finding Consistent Clusters in Data Partitions

Abstract: Given an arbitrary data set, to which no particular parametrical, statistical or geometrical structure can be assumed, different clustering algorithms will in general produce different data partitions. In fact, several partitions can also be obtained by using a single clustering algorithm due to dependencies on initialization or the selection of the value of some design parameter. This paper addresses the problem of finding consistent clusters in data partitions, proposing the analysis of the most common assoc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
142
0
7

Year Published

2005
2005
2021
2021

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 224 publications
(151 citation statements)
references
References 17 publications
(19 reference statements)
0
142
0
7
Order By: Relevance
“…When the single linkage method has been used for consensus clustering it has been referred to as the majority rule or the quota rule 21,41 . The DIRECT procedure in CLUTO provided the next consensus clustering method.…”
Section: Consensus Clustering Methodsmentioning
confidence: 99%
“…When the single linkage method has been used for consensus clustering it has been referred to as the majority rule or the quota rule 21,41 . The DIRECT procedure in CLUTO provided the next consensus clustering method.…”
Section: Consensus Clustering Methodsmentioning
confidence: 99%
“…Fred and Jain [9,11,12,10] suggest to use the k-means clustering algorithm several times with random initial conditions. In each of the clustering trials, the number of clusters, k, is either fixed or chosen randomly in the range k ∈ [k min , k max ].…”
Section: Consensus Clusteringmentioning
confidence: 99%
“…Each value of k is used to accumulate evidence about the clustering structure using two different approaches: -In the first approach, what we will refer to as a consensus matrix, is computed. This entails simply counting for each k whether or not pairs of data points in the data set belong to the same basin of attraction (mode), for then to compute the average over all k. Based on the consensus matrix, a hierarchical clustering approach similar to that used in [9] and [11] is utilized in order to obtain the final clustering result. -The second approach we investigate, is based on for each k to compute an information theoretic divergence measure between pairs of modes resulting in a similarity matrix between modes, for then to average over all k. Then, a spectral clustering procedure is executed on this matrix, similar to [1].…”
Section: Introductionmentioning
confidence: 99%
“…A popular technique for merging is called 'majority voting' [11,12] which is a pair-counting method extended over multiple clusterings. Using a co-association matrix of data points, where pairs of points are given a score if they appear in the same cluster over all available clusterings.…”
Section: Related Workmentioning
confidence: 99%