Progress in Artificial Intelligence
DOI: 10.1007/978-3-540-77002-2_33
|View full text |Cite
|
Sign up to set email alerts
|

Experiments for the Number of Clusters in K-Means

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 28 publications
(24 citation statements)
references
References 14 publications
0
24
0
Order By: Relevance
“…If the result is 10 or greater, k + 1 clusters are preferable. Chiang and Mirkin (2007) reported experimental results supporting Hartigan's as the method that produces the most accurate number of clusters. An important issue was whether the same number of clusters was reached in each university sample.…”
Section: Discussionmentioning
confidence: 48%
“…If the result is 10 or greater, k + 1 clusters are preferable. Chiang and Mirkin (2007) reported experimental results supporting Hartigan's as the method that produces the most accurate number of clusters. An important issue was whether the same number of clusters was reached in each university sample.…”
Section: Discussionmentioning
confidence: 48%
“…have a high level of dynamism, the need to specify a priori the number of clusters is an important drawback of this clustering method. For this reason, the proposed solution uses dynamic clustering [6], which classifies learners on the basis of similar learning needs and interests, without requiring an initial indication of the number of clusters. In particular, the proposed approach uses the Silhouette index [5] to estimate the optimal number of clusters in which to group the data set and the K-means algorithm [4] to cluster the data set into the optimal, previously defined partition.…”
Section: The Dynamic Clustering Of Learnersmentioning
confidence: 99%
“…The algorithm calculates the score of all possible number of clusters, provided in the range, and calculate their score using Bayesian Information Criterion (BIC) Akaike Information Criteria (AIC), and the one with the best score is output. There are in fact other kmeans variants that also attempt to find the right number of clusters by follow up analysis, but it has been experimentally proven in [12] that those do not provide results as consistent as iKMeans. Evidently the fact that iKMeans provides better results than other algorithms does not mean that it always provides good results.…”
Section: Kmeans and Ikmeansmentioning
confidence: 99%
“…Their cluster determined through the distances between the entities and the only other seed, the center itself. At the end, small clusters are removed according to a threshold and the final seeds are then used as initial seeds in the KMeans Algorithm, iKMeans has also been successfully applied to a number of different comparative experiments, such as in [12] The importance of finding good seeds has been object of much research, algorithms such as the kmeans++, introduced by [11] have already attempted this problem, but this for instance does not find the number of clusters, just better seeds for the given number of clusters. Xmeans, introduced by [8] is an example of an algorithm that in fact tries to determine the right number of clusters, but in this algorithm a range where the true number of clusters is has to be provided.…”
Section: Kmeans and Ikmeansmentioning
confidence: 99%
See 1 more Smart Citation