1979
DOI: 10.2307/2346830
|View full text |Cite
|
Sign up to set email alerts
|

Algorithm AS 136: A K-Means Clustering Algorithm

Abstract: JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact support@jstor.org.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

14
6,007
1
108

Year Published

2004
2004
2024
2024

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 9,902 publications
(6,484 citation statements)
references
References 3 publications
14
6,007
1
108
Order By: Relevance
“…The clustering algorithm we use is the K-means algorithm, as implemented by Hartigan and Wong. 29,30 A particularly salient feature of this algorithm, for our purposes, is that the computational expense scales linearly with the number of items (loops) to cluster. However, unlike clustering algorithms that calculate similarity between every pair of items (and thus scale quadratically), the number of clusters must be specified in advance for K-means.…”
Section: Clusteringmentioning
confidence: 99%
“…The clustering algorithm we use is the K-means algorithm, as implemented by Hartigan and Wong. 29,30 A particularly salient feature of this algorithm, for our purposes, is that the computational expense scales linearly with the number of items (loops) to cluster. However, unlike clustering algorithms that calculate similarity between every pair of items (and thus scale quadratically), the number of clusters must be specified in advance for K-means.…”
Section: Clusteringmentioning
confidence: 99%
“…In addition, each cluster is represented by a centre point with properties which are assumed to be representative of all catchments in the cluster. The k-means clustering algorithm (Hartigan and Wong, 1979) described in the next section is used to assign the catchments into nc clusters and also to calculate a centre vector of the spatial variables for each cluster. In addition, a standard deviation vector for each cluster can be calculated using the resulting centre vector and the vectors of the spatial data for all catchments in the cluster.…”
Section: New Neuro-fuzzy National P Export Modelmentioning
confidence: 99%
“…Some defined distance measure such as the Euclidean distance is often used to determine proximity of the data in a cluster. The k-means clustering algorithm (Hartigan and Wong, 1979) is one of the simplest unsupervised learning algorithms for this partitioning when the number of clusters (k) is known a priori. The number of clusters is normally determined based on the amount and characteristics of the data which is used in calibrating the model.…”
Section: K-means Clustering Algorithmmentioning
confidence: 99%
“…K-mean clustering is a statistical method that partitions a given data set into a specific number of clusters in which each data belongs to the cluster with the nearest mean. For a more detailed algorithm, refer to (Hartigan and Wong 1979).…”
Section: K-mean Clustering Algorithmmentioning
confidence: 99%