2017
DOI: 10.1007/978-3-319-57454-7_8
|View full text |Cite
|
Sign up to set email alerts
|

Feature Ranking of Large, Robust, and Weighted Clustering Result

Abstract: Abstract. A clustering result needs to be interpreted and evaluated for knowledge discovery. When clustered data represents a sample from a population with known sample-to-population alignment weights, both the clustering and the evaluation techniques need to take this into account. The purpose of this article is to advance the automatic knowledge discovery from a robust clustering result on the population level. For this purpose, we derive a novel ranking method by generalizing the computation of the Kruskal-… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
3
3
2

Relationship

6
2

Authors

Journals

citations
Cited by 9 publications
(10 citation statements)
references
References 17 publications
(25 reference statements)
0
7
0
Order By: Relevance
“…The SSE values were computed with formula (1) for the whole data. Finally, statistical comparison between the methods was performed with the nonparametric Kruskal-Wallis test [48,49], since in most cases the clustering errors were not normally distributed. The significance level was set to 0.05.…”
Section: Methodsmentioning
confidence: 99%
“…The SSE values were computed with formula (1) for the whole data. Finally, statistical comparison between the methods was performed with the nonparametric Kruskal-Wallis test [48,49], since in most cases the clustering errors were not normally distributed. The significance level was set to 0.05.…”
Section: Methodsmentioning
confidence: 99%
“…These benchmark sets are synthetic datasets, suggested for use when testing any algorithm dealing with clustering spherical or Gaussian data. Here, we restrict ourselves to the benchmarks with at most 20 clusters, because the interpretation and knowledge discovery from the clustering results with a large number of prototypes might become tedious [6,49]. Therefore, the number of clusters was also tested with K = 2 − 25.…”
Section: Methodsmentioning
confidence: 99%
“…Hierarchical clustering constructs a tree structure from data to present layers of clustering results, but because of the pairwise distance matrix requirement, the basic form of the method is not scalable to a large volume of data [5]. Moreover, many clustering algorithms, including hierarchical clustering, can produce clusters of arbitrary shapes in the data space, which might be difficult to interpret for knowledge discovery [6].…”
Section: Introductionmentioning
confidence: 99%
“…The SSE values were computed with formula (1) for the whole data. Finally, statistical comparison between the methods was performed with the nonparametric Kruskal-Wallis test [37], [38], since in most of the cases the clustering errors were not normally distributed. The significance level was set to 0.05.…”
Section: Methodsmentioning
confidence: 99%