2013
DOI: 10.1016/j.patcog.2012.10.021
|View full text |Cite
|
Sign up to set email alerts
|

How large should ensembles of classifiers be?

Abstract: Esta es la versión de autor del artículo publicado en: This is an author produced version of a paper published in: AbstractWe propose to determine the size of a parallel ensemble by estimating the minimum number of classifiers that are required to obtain stable aggregate predictions. Assuming that majority voting is used, a statistical description of the convergence of the ensemble prediction to its asymptotic (infinite size) limit is given. The analysis of the voting process shows that for most test instance… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
28
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 55 publications
(29 citation statements)
references
References 27 publications
1
28
0
Order By: Relevance
“…The results from [13,14] are not equally clear-cut and they indicate that different machine learning problems need different ensemble sizes. Both papers investigate how to dynamically create a dynamic stopping criterion to decide when to stop adding more classifiers to an ensemble.…”
Section: B Ensemble Sizementioning
confidence: 91%
“…The results from [13,14] are not equally clear-cut and they indicate that different machine learning problems need different ensemble sizes. Both papers investigate how to dynamically create a dynamic stopping criterion to decide when to stop adding more classifiers to an ensemble.…”
Section: B Ensemble Sizementioning
confidence: 91%
“…In another direction, the paper [13] studied algorithmic convergence in terms of a different criterion, namely the "disagreement probability" δ t := P(M t (X) = M ∞ (X) D), where M ∞ (X) is the infinite ensemble analogue of the majority vote M t (X). In that work, an informal derivation is given to show that δ t is of order O( 1 √ t ).…”
Section: Related Workmentioning
confidence: 99%
“…[5,6,7,8,9,10,11] among others), comparatively little is known about how the prediction error depends on the number of classifiers (ensemble size). In particular, only a handful of works have considered this question from a theoretical standpoint [12,13,1,2] (cf. Section 1.2).…”
Section: Introductionmentioning
confidence: 99%
“…In the setting of classification, much of the literature has studied convergence in terms of the misclassification probability for majority voting, denoted err t (a counterpart of mse t ), which is viewed as a random variable that depends on ξ t and D. For this measure of error, the convergence of E[err t |D] and var(err t |D) as t → ∞ has been analyzed in the papers (Ng and Jordan, 2001;Lopes, 2016;Cannings and Samworth, 2017), which have developed asymptotic formulas for E[err t |D], as well as bounds for var(err t |D). Related results for a different measure of error can also be found in Hernández-Lobato et al (2013). More recently, our companion paper (Lopes, 2019) has developed a bootstrap method for measuring the convergence of err t , which is able to circumvent some of the limitations of analytical results.…”
Section: Related Work and Contributionsmentioning
confidence: 99%