2006
DOI: 10.1016/j.inffus.2005.01.008
|View full text |Cite
|
Sign up to set email alerts
|

Moderate diversity for better cluster ensembles

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
110
0
9

Year Published

2006
2006
2018
2018

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 188 publications
(127 citation statements)
references
References 25 publications
4
110
0
9
Order By: Relevance
“…Several studies focused on understanding how diversity was handled on various ensemble creation techniques like AdaBoost or Bagging [11,12]. Finally, many techniques have been proposed for exploiting diversity for finding good ensembles [13][14][15][16][17][18]. It was even proposed to voluntarily overtrain the classifiers in order to create diversity between them [19].…”
Section: Diversity In Ensembles Of Classifiersmentioning
confidence: 99%
See 1 more Smart Citation
“…Several studies focused on understanding how diversity was handled on various ensemble creation techniques like AdaBoost or Bagging [11,12]. Finally, many techniques have been proposed for exploiting diversity for finding good ensembles [13][14][15][16][17][18]. It was even proposed to voluntarily overtrain the classifiers in order to create diversity between them [19].…”
Section: Diversity In Ensembles Of Classifiersmentioning
confidence: 99%
“…Kuncheva also reported in [6] that the improvement on the best individual accuracy by forcing diversity is negligible. In [14], Hadjitodorov showed that, in some particular cases, using moderate diversity can produce better ensemble than maximum measure of diversity.…”
Section: Limits Of Diversity Measuresmentioning
confidence: 99%
“…In this section we introduce the pairwise similarity matrices between examples [28,45], since they are used to compute the stability measures proposed in this paper. The similarity matrix is a sort of distributed memory of the clusters, by which memberships of pairs of examples to the same cluster are stored.…”
Section: Similarity Matrixmentioning
confidence: 99%
“…There are two factors that influence the performance of this approach: one is the accuracy of the individual clusters (P i ) and the other is the diversity within the ensemble E. Accuracy is maintained by tuning a set of effective clustering methods to obtain the best set of results. Regarding the diversity of E, it was shown in [15] that a moderate level of dissimilarity among the ensemble members (E) improves the consensus results. For this, we studied the diversity within E, using the Rand Index (RI) similarity measure [16], and created a more effective sub-set of cluster solutions to represent the new ensemble, denoted here as E .…”
Section: Consensus Clustering (Cc) Frameworkmentioning
confidence: 99%