2019
DOI: 10.3390/e21070683
|View full text |Cite
|
Sign up to set email alerts
|

An Entropy Regularization k-Means Algorithm with a New Measure of between-Cluster Distance in Subspace Clustering

Abstract: Although within-cluster information is commonly used in most clustering approaches, other important information such as between-cluster information is rarely considered in some cases. Hence, in this study, we propose a new novel measure of between-cluster distance in subspace, which is to maximize the distance between the center of a cluster and the points that do not belong to this cluster. Based on this idea, we firstly design an optimization objective function integrating the between-cluster distance and en… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
3
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 32 publications
0
3
0
Order By: Relevance
“…We have also compared our metrics to some related work. We have used as a comparison the following results from [36]: KMEA, WKME, EWKM, ESSC, AFKM, SC, SSC-MP, ERKM; and from [29]: Bayes Network Classifier, J48, Random Forest, OneR. In Table 2 we can see that the F1-Score for the two ABARC cases is better than all of the others, but the Kappa Score is better only after removing hybrids.…”
Section: Metricsmentioning
confidence: 99%
See 1 more Smart Citation
“…We have also compared our metrics to some related work. We have used as a comparison the following results from [36]: KMEA, WKME, EWKM, ESSC, AFKM, SC, SSC-MP, ERKM; and from [29]: Bayes Network Classifier, J48, Random Forest, OneR. In Table 2 we can see that the F1-Score for the two ABARC cases is better than all of the others, but the Kappa Score is better only after removing hybrids.…”
Section: Metricsmentioning
confidence: 99%
“…Precision Recall F1-Score Kappa Score KMEA [23] 81.2 WKME [13] 79.8 EWKM [15] 82.6 ESSC [6] 84.8 AFKM [1] 81.6 SC [31] 47.2 SSC-MP [32] 76.7 ERKM [36] 90. Seeds dataset.…”
Section: Algorithmmentioning
confidence: 99%
“…r distances, KICIC), which is able to integrate intra-cluster solve the clustering problem of high-dimensional data [8]. rithm based on distance threshold and weighted sample.…”
Section: Traditional K-means Algorithmmentioning
confidence: 99%
“…In each iteration of clustering, it computes the optimal weight of attributes according to the change of centroid vector which minimizes the sum of distance between each instance and the centroid [28]. Xiaohui proposed a novel K-Means type method (a weighting K-Means clustering approach by integrating intra-cluster and inter-cluster distances, KICIC), which is able to integrate intra-cluster compactness and inter-cluster separation to solve the clustering problem of high-dimensional data [8]. Jiyong proposed a K-Means clustering algorithm based on distance threshold and weighted sample.…”
Section: Introductionmentioning
confidence: 99%
“…Another approach for feature weighting in centre‐based clustering is obtained by incorporating a Shannon entropic regularizer on feature weights (Jing, Ng, & Huang, 2007). Some relevant works on the endorsement of using an entropy regularizer include Chakraborty, Paul, Das, and Xu (2020), Pihur, Datta, and Datta (2007), Xiong, Wang, Huang, and Zeng (2019), and Zhou, Chen, Chen, Zhang, and Li (2016). A comprehensive overview of the available clustering techniques for high‐dimensional data in the form of R package ‘HDclassif’ can be found in Bergé, Bouveyron, and Girard (2012).…”
Section: Introductionmentioning
confidence: 99%