1996
DOI: 10.1016/0167-8655(95)00119-0
|View full text |Cite
|
Sign up to set email alerts
|

New methods for the initialisation of clusters

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
27
0

Year Published

2002
2002
2020
2020

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 64 publications
(27 citation statements)
references
References 9 publications
0
27
0
Order By: Relevance
“…At the same time, the selected values have to be significantly smaller than the number of objects in the data sets, which is the main motivation for performing data clustering. Reported studies [2][3][4][5][6][7][8][9][10][11][12][13][14][15][16][17][18] on K-means clustering and its applications usually do not contain any explanation or justification for selecting particular values for K. Table 1 lists the numbers of clusters and objects and the corresponding data sets used in those studies. Two observations could be made when analysing the data in the table.…”
Section: Values Of K Specified Within a Range Or Setmentioning
confidence: 99%
“…At the same time, the selected values have to be significantly smaller than the number of objects in the data sets, which is the main motivation for performing data clustering. Reported studies [2][3][4][5][6][7][8][9][10][11][12][13][14][15][16][17][18] on K-means clustering and its applications usually do not contain any explanation or justification for selecting particular values for K. Table 1 lists the numbers of clusters and objects and the corresponding data sets used in those studies. Two observations could be made when analysing the data in the table.…”
Section: Values Of K Specified Within a Range Or Setmentioning
confidence: 99%
“…Methods for automatic initialization of clusters have been proposed in the literature [10 -12]. Al-Daoud and Roberts [10] proposed two methods, the first of which picks points randomly in evenly spaced cells across the entire histogram of the data and reduces the number until the required seeds are found. The second method tries to optimize the sum of squares of the distances from the cluster centres.…”
Section: Introductionmentioning
confidence: 99%
“…The assignment should include a vector quantization request; Katsavounidis variant obtains the point with the most Euclidean standard as the first center. AlDaoud's density-based method first regularly partitions the data space into M decomposed hyper-cubes [27]. Then this randomly selects K Nm/N points of hypercube m (m {1,2,…,M}) to create a number of K centers where Nm is the number of points in hypercube m. Bradley and Fayyad's method [28] begins by randomly partitioning the data set into J subsets.…”
Section: IImentioning
confidence: 99%