2009
DOI: 10.1214/08-ejs196
|View full text |Cite
|
Sign up to set email alerts
|

Model selection by resampling penalization

Abstract: In this paper, a new family of resampling-based penalization procedures for model selection is defined in a general framework. It generalizes several methods, including Efron's bootstrap penalization and the leave-one-out penalization recently proposed by Arlot (2008), to any exchangeable weighted bootstrap resampling scheme. In the heteroscedastic regression framework, assuming the models to have a particular structure, these resampling penalties are proved to satisfy a non-asymptotic oracle inequality with l… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
64
0

Year Published

2010
2010
2021
2021

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 35 publications
(67 citation statements)
references
References 68 publications
3
64
0
Order By: Relevance
“…In particular, precise concentration inequalities for R Lpo p (m) would be needed. According to the assumptions under which similar results have been obtained for polynomial collections of models (Arlot 2008(Arlot , 2009Celisse 2008b), we can conjecture that (at least) the following assumptions would be required to derive an upper bound on s − s m Lpo p (D) 2 n : • moment inequalities for the errors ε i : ∀q ≥ 2, ε q ≤ Cq β for some C, β > 0 (for instance, ε sub-Gaussian), • a uniform upper bound on σ (see (BN) in Proposition 1), and probably a uniform lower bound or mild smoothness assumptions on σ (·), • a relationship between the minimum number of points in the intervals of a segmentation and the smoothness of the regression function s. Note that assumption (BV) in Proposition 1 entails that for every λ ∈ m and t i ∈ λ,…”
Section: Theoretical Guaranteesmentioning
confidence: 65%
See 2 more Smart Citations
“…In particular, precise concentration inequalities for R Lpo p (m) would be needed. According to the assumptions under which similar results have been obtained for polynomial collections of models (Arlot 2008(Arlot , 2009Celisse 2008b), we can conjecture that (at least) the following assumptions would be required to derive an upper bound on s − s m Lpo p (D) 2 n : • moment inequalities for the errors ε i : ∀q ≥ 2, ε q ≤ Cq β for some C, β > 0 (for instance, ε sub-Gaussian), • a uniform upper bound on σ (see (BN) in Proposition 1), and probably a uniform lower bound or mild smoothness assumptions on σ (·), • a relationship between the minimum number of points in the intervals of a segmentation and the smoothness of the regression function s. Note that assumption (BV) in Proposition 1 entails that for every λ ∈ m and t i ∈ λ,…”
Section: Theoretical Guaranteesmentioning
confidence: 65%
“…In particular, Loo is asymptotically equivalent to AIC or C p in several frameworks where they are asymptotically optimal, and other CV methods have similar performances, provided the size of the training sample is close enough to the sample size (see for instance Li 1987;Shao 1997;Dudoit and van der Laan 2005). In addition, CV methods are robust to heteroscedasticity of data (Arlot 2008), as well as several other resampling methods (Arlot 2009). Therefore, CV is a natural alternative to penalization procedures assuming homoscedasticity.…”
Section: Cross-validationmentioning
confidence: 99%
See 1 more Smart Citation
“…However, complexity-based penalties cannot be used straightforwardly since further studies are needed to calibrate the numerical constants involved. We refer to a recent paper by Arlot (2009) where resampling strategies are depicted in the classification setup in order to address this issue. The simulation study we performed reveals the potential but also the limitations of these scoring algorithms.…”
Section: Resultsmentioning
confidence: 99%
“…We ran K-means clustering [Bis95] on the free-form chunk dataset, to cluster the chunks according to their motion (see Cluster attack). We applied the v-fold cross-validation algorithm [Arl07] to determine the optimal number of clusters in our dataset. The outcome was K = 6.…”
Section: Attack Datasetsmentioning
confidence: 99%