2004
DOI: 10.1214/aos/1079120131
|View full text |Cite
|
Sign up to set email alerts
|

Optimal aggregation of classifiers in statistical learning

Abstract: Classification can be considered as nonparametric estimation of sets, where the risk is defined by means of a specific distance between sets associated with misclassification error. It is shown that the rates of convergence of classifiers depend on two parameters: the complexity of the class of candidate sets and the margin parameter. The dependence is explicitly given, indicating that optimal fast rates approaching O(n −1 ) can be attained, where n is the sample size, and that the proposed classifiers have th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

26
524
0

Year Published

2006
2006
2023
2023

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 461 publications
(550 citation statements)
references
References 28 publications
26
524
0
Order By: Relevance
“…Lugosi (2002) and Devroye and Lugosi (1995) then confirmed these reservations by studying the interpolation case where the best classification error inf t∈S P[t(X ) = Y ] of a given class S is nonzero but small. By further analyzing the problem, Mammen and Tsybakov (1999), Tsybakov (2004) and Massart and Nédélec (2005) show that the behavior of the regression function η : x → P[Y = 1|X = x] around 1/2 is crucial. They indeed introduce some margin conditions that can be written in the following general way:…”
Section: Resultsmentioning
confidence: 99%
“…Lugosi (2002) and Devroye and Lugosi (1995) then confirmed these reservations by studying the interpolation case where the best classification error inf t∈S P[t(X ) = Y ] of a given class S is nonzero but small. By further analyzing the problem, Mammen and Tsybakov (1999), Tsybakov (2004) and Massart and Nédélec (2005) show that the behavior of the regression function η : x → P[Y = 1|X = x] around 1/2 is crucial. They indeed introduce some margin conditions that can be written in the following general way:…”
Section: Resultsmentioning
confidence: 99%
“…First, we recall the definition of the margin assumption introduced in [30]. Margin Assumption(MA): The probability measure π satisfies the margin assumption MA(κ), where κ ≥ 1 if we have…”
Section: Suboptimality Of Penalized Erm Proceduresmentioning
confidence: 99%
“…The assumption that the noise rates are not equal to 1/2 can be relaxed (at the cost of error values no longer approaching zero) if we assume the weight of the area with noise rate close to 1/2 is bounded (e.g., by applying Tsybakov's noise condition [15]). …”
Section: Noise Rates Different From 1/2mentioning
confidence: 99%