2011
DOI: 10.1016/j.patrec.2010.11.006
|View full text |Cite
|
Sign up to set email alerts
|

Towards a standard methodology to evaluate internal cluster validity indices

Abstract: The evaluation and comparison of internal cluster validity indices is a critical problem in the clustering area. The methodology used in most of the evaluations assumes that the clustering algorithms work correctly. We propose an alternative methodology that does not make this often false assumption. We compared 7 internal cluster validity indices with both methodologies and concluded that the results obtained with the proposed methodology are more representative of the actual capabilities of the compared indi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
27
0

Year Published

2011
2011
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 57 publications
(28 citation statements)
references
References 28 publications
1
27
0
Order By: Relevance
“…Consequently, the smaller their values the better is the solution. This is exactly what Deviation and Connectivity measures do respectively (Handl and Knowles, 2007;Hruschka et al, 2009;Gurrutxaga et al, 2011).…”
Section: Deviation and Connectivitysupporting
confidence: 54%
See 1 more Smart Citation
“…Consequently, the smaller their values the better is the solution. This is exactly what Deviation and Connectivity measures do respectively (Handl and Knowles, 2007;Hruschka et al, 2009;Gurrutxaga et al, 2011).…”
Section: Deviation and Connectivitysupporting
confidence: 54%
“…These facts often hinders the data analysis step because experts must evaluate all the different solutions generated by the algorithm, which is highly time consuming and quite arbitrary because the selection will depend on the subjectivity of the expert due to the fact that all of them are potentially valid. For this reason, the application of evaluation functions for automatically scoring the clustering solutions has become the key for helping experts to select the best (Gurrutxaga et al, 2011). These evaluation functions define metrics that measure the cluster quality by using the same features included in the data set.…”
Section: Looking For the Most Suitable Patternsmentioning
confidence: 99%
“…VIC was tested using 50 different data sets, where it significantly outperforms other well known cluster indexes (Rodríguez et al, 2018). Unlike other internal indexes, which tend to prefer clusters with specific shapes, such as hyperspheres (Lago-Fernández & Corbacho, 2010;Halkidi & Vazirgiannis, 2008;Gurrutxaga et al, 2011), VIC does not assume a specific shape. Similarly, other indexes tend to prefer higher number of clusters (Dubes, 1987), whereas VIC does not.…”
Section: Validity Index Using Supervised Classifiersmentioning
confidence: 99%
“…Recently, it is suggested by [4] that internal evaluation measures should be evaluated with external indices. They used Kmeans to generate different partitions and compared the best partitions by internal measures to the best partitions by external measures (real partitions).…”
Section: Related Workmentioning
confidence: 99%
“…[3]). However, evaluation of cluster models is a difficult task because of the unsupervised nature of clustering [4].…”
Section: Introductionmentioning
confidence: 99%