2019
DOI: 10.1007/s10208-019-09427-x
|View full text |Cite
|
Sign up to set email alerts
|

Mean Estimation and Regression Under Heavy-Tailed Distributions: A Survey

Abstract: We survey some of the recent advances in mean estimation and regression function estimation. In particular, we describe sub-Gaussian mean estimators for possibly heavy-tailed data both in the univariate and multivariate settings. We focus on estimators based on median-of-means techniques but other methods such as the trimmed mean and Catoni's estimator are also reviewed. We give detailed proofs for the cornerstone results. We dedicate a section on statistical learning problems-in particular, regression functio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
124
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
9
1

Relationship

0
10

Authors

Journals

citations
Cited by 123 publications
(125 citation statements)
references
References 73 publications
1
124
0
Order By: Relevance
“…To achieve robustness in both feature and response spaces, recent years have witnessed a rapid development of the "median-of-means" (MOM) principle, which dates back to [40] and [22], and a variety of MOM-based procedures for regression and classification in both low-an high-dimensional settings [12,13,26,27,33,35]. We refer to [34] for a recent survey. An interesting open problem is how to efficiently incorporate the MOM principle with nonconvex regularization or iteratively reweighted 1 -regularization so as to achieve high degree of robustness and variable selection consistency simultaneously.…”
Section: Related Literaturementioning
confidence: 99%
“…To achieve robustness in both feature and response spaces, recent years have witnessed a rapid development of the "median-of-means" (MOM) principle, which dates back to [40] and [22], and a variety of MOM-based procedures for regression and classification in both low-an high-dimensional settings [12,13,26,27,33,35]. We refer to [34] for a recent survey. An interesting open problem is how to efficiently incorporate the MOM principle with nonconvex regularization or iteratively reweighted 1 -regularization so as to achieve high degree of robustness and variable selection consistency simultaneously.…”
Section: Related Literaturementioning
confidence: 99%
“…Robust statistics substantially broaden the capacity of traditional data models so that they can embrace real-world examples. For instance, a recent series of works such as Catoni (2012); Minsker (2015); Devroye et al (2016); Fan et al (2017; Lugosi and Mendelson (2019) study how to handle heavy-tailed data in point estimation and regression analysis. These works do not assume any parametric form of the data distribution but a bounded moment, and they show that the estimator based on median-of-means or robustified risk minimization exhibits sub-Gaussian behavior around the truth.…”
Section: Robust Statisticsmentioning
confidence: 99%
“…where β > 0 is a parameter to be tuned and ϕ is non-decreasing and called influence function. The deviation performance of this estimator is much better than X. Catoni's idea has been broadly applied to many research problems, see for instance [1,15,5,6,7,11,12,17]. The finite variance assumption plays an important role in Catoni's analysis, but it rules out many interesting distributions such as Pareto law [10,16,4,8], which describes the distributions of wealth and social networks.…”
Section: Introductionmentioning
confidence: 99%