The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2013
DOI: 10.3906/elk-1010-869
|View full text |Cite
|
Sign up to set email alerts
|

K-means algorithm with a novel distance measure

Abstract: Abstract:In this paper, we describe an essential problem in data clustering and present some solutions for it. We investigated using distance measures other than Euclidean type for improving the performance of clustering. We also developed an improved point symmetry-based distance measure and proved its efficiency. We developed a k-means algorithm with a novel distance measure that improves the performance of the classical k-means algorithm. The proposed algorithm does not have the worst-case bound on running … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 11 publications
0
4
0
Order By: Relevance
“…Word2vec [34] then mapped the candidate words into the vector space. The K-means algorithm [35] was used to cluster the candidate vectors to obtain positive and negative clustering centers. The candidate words were individually determined positive or negative when the corresponding candidate vectors were close to the positive and negative clustering centers.…”
Section: Data Preprocessingmentioning
confidence: 99%
“…Word2vec [34] then mapped the candidate words into the vector space. The K-means algorithm [35] was used to cluster the candidate vectors to obtain positive and negative clustering centers. The candidate words were individually determined positive or negative when the corresponding candidate vectors were close to the positive and negative clustering centers.…”
Section: Data Preprocessingmentioning
confidence: 99%
“…(A4) In present time various distance measure is available for clustering and these distance measure groups under Minkowski, L(1), L(2), Inner product, Shannon's entropy, Combination, Intersection and Fidelity family [4][14] [35]. In this section, the paper describes various distance measures under these families [4][5][6][7][8][9][10][11][12][13][14][15][16][17][18] [32][33][34][35].…”
Section: Distance Measures Taxonomymentioning
confidence: 99%
“…Distance measures are not only essential to solve the clustering problem, but it is also solved to pattern recognition, classification, retrieval related problems [4], help to the derivation of new distance measure [5], text classification and clustering [6], document content comparison [7], time-series data management [8], uncertain data classification [9] and clustering [l0], bio-cryptic authentication in cloud databases [11], spatial concentration [12], location fingerprinting [13], author profiling [14], combining density [15], heavy aggregation operators [16], analyzing inconsistent information [17], network intrusion anomaly detection [18] for high volume, variety and velocity. The objective of this paper is identifying the best cluster distance measure for cluster creation in the big data mining and this objective is obtained by the six sections.…”
Section: Introductionmentioning
confidence: 99%
“…It separates a data set into subsets or clusters so that data values in the same cluster have some common characteristics or attributes [2]. It aims to divide the data into groups (clusters) of similar objects [3]. The objects in the same cluster are more identical to each other than to those in other clusters.…”
Section: Introductionmentioning
confidence: 99%