2021
DOI: 10.48550/arxiv.2111.11694
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MARS via LASSO

Abstract: MARS is a popular method for nonparametric regression introduced by Friedman in 1991. MARS fits simple nonlinear and non-additive functions to regression data. We propose and study a natural LASSO variant of the MARS method. Our method is based on least squares estimation over a convex class of functions obtained by considering infinite-dimensional linear combinations of functions in the MARS basis and imposing a variation based complexity constraint. We show that our estimator can be computed via finite-dimen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 31 publications
(135 reference statements)
0
3
0
Order By: Relevance
“…Our CV framework is based on a general principle and should be looked upon as providing a general recipe to develop theoretically tractable CV versions of potentially any other estimator which uses a tuning parameter. For example, using our framework, one should be able to develop theoretically tractable CV versions of the Total Variation Denoising estimator proposed by Rudin et al (1992) (also see Hütter and Rigollet (2016), Sadhanala et al (2016), Chatterjee and Goswami (2019b)), the Hardy Krauss estimator (see Fang et al (2021), Ortelli and van de Geer (2020)), the Optimal Regression Tree estimator proposed in Chatterjee and Goswami (2019a), a higher dimensional version of Trend Filtering of order 2 proposed in Ki et al (2021) and potentially many more. As a start, the Zero Doubling method of constructing completion estimators alongwith using a geometrically doubling grid of candidate tuning values should be useable in these problems.…”
Section: Other Signal Denoising Methodsmentioning
confidence: 99%
“…Our CV framework is based on a general principle and should be looked upon as providing a general recipe to develop theoretically tractable CV versions of potentially any other estimator which uses a tuning parameter. For example, using our framework, one should be able to develop theoretically tractable CV versions of the Total Variation Denoising estimator proposed by Rudin et al (1992) (also see Hütter and Rigollet (2016), Sadhanala et al (2016), Chatterjee and Goswami (2019b)), the Hardy Krauss estimator (see Fang et al (2021), Ortelli and van de Geer (2020)), the Optimal Regression Tree estimator proposed in Chatterjee and Goswami (2019a), a higher dimensional version of Trend Filtering of order 2 proposed in Ki et al (2021) and potentially many more. As a start, the Zero Doubling method of constructing completion estimators alongwith using a geometrically doubling grid of candidate tuning values should be useable in these problems.…”
Section: Other Signal Denoising Methodsmentioning
confidence: 99%
“…Other models for multivariate TV. There has been a recent surge of work studying different generalizations of TV and trend filtering penalties to multiple dimensions, for example, Bibaut and van der Laan (2019); ; Ortelli and van de Geer (2021); Chatterjee and Goswami (2021); Ki et al (2021). Many of these works are based on the notion of Hardy-Krause variation of a multivariate function, or the related notion of Vitali variation.…”
Section: Continuous-time Multivariate Tv Methodsmentioning
confidence: 99%
“…However, for many function classes, this kind of approximability may not hold. For example, we can consider the class of Hardy Krause Bounded Variation Functions (see Fang et al (2021)) or its higher order versions (see Ki et al (2021)) where the existing covering argument produces nets (to estimate metric entropy) which are not necessarily rectangular piecewise constant/linear respectively. These function classes are also known not to suffer from the curse of dimensionality in the sense that the metric entropy does not grow exponentially in 1 with the dimension d. More generally, it would be very interesting to come up with computationally efficient and statistically rate optimal online prediction algorithms for such function classes.…”
Section: 2mentioning
confidence: 99%