2018
DOI: 10.1080/01621459.2016.1273116
|View full text |Cite
|
Sign up to set email alerts
|

Boosting in the Presence of Outliers: Adaptive Classification With Nonconvex Loss Functions

Abstract: This paper examines the role and efficiency of the non-convex loss functions for binary classification problems. In particular, we investigate how to design a simple and effective boosting algorithm that is robust to the outliers in the data. The analysis of the role of a particular non-convex loss for prediction accuracy varies depending on the diminishing tail properties of the gradient of the loss -the ability of the loss to efficiently adapt to the outlying data, the local convex properties of the loss and… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 28 publications
(15 citation statements)
references
References 53 publications
0
15
0
Order By: Relevance
“…For the number of cases and deaths, we used the Poisson distribution and for the growth rate, we used the Gaussian distribution. We used natural logarithm transformation (variables: Dens, GDP, Time, Tour) because gradient boosting regression may produce biased results in the presence of outliers [ 44 ]. The fitted model was then used to make predictions on the test data.…”
Section: Methodsmentioning
confidence: 99%
“…For the number of cases and deaths, we used the Poisson distribution and for the growth rate, we used the Gaussian distribution. We used natural logarithm transformation (variables: Dens, GDP, Time, Tour) because gradient boosting regression may produce biased results in the presence of outliers [ 44 ]. The fitted model was then used to make predictions on the test data.…”
Section: Methodsmentioning
confidence: 99%
“…There are also many other statistical problems in which robust estimation has been recently revisited from the point of view of minimax rates. This includes scale and covariance matrix estimation (Chen et al, 2018;Comminges et al, 2018), matrix completion (Klopp et al, 2017), multivariate regression (Dalalyan and Thompson, 2019;Gao, 2020;Geoffrey, 2019), classification (Cannings et al, 2018;Li and Bradic, 2018), subspace clustering (Soltanolkotabi and Candès, 2012), community detection (Cai and Li, 2015), etc. Properties of robust M -estimators in high-dimensional settings are studied in (Elsener and van de Geer, 2018;Loh, 2017).…”
Section: Factmentioning
confidence: 99%
“…Shi et al ( 2010) used gene expressions to predict breast cancer clinical status with 164 estrogen receptor positive cases and 114 negative cases. The same data set has been evaluated for a variety of robust classification algorithms (Li and Bradic, 2018;Wang, 2018a,b). The analysis described below is reproduced in a vignette distributed with the R package mpath.…”
Section: Predicting Breast Cancer Clinical Statusmentioning
confidence: 99%
“…However, a breakdown point analysis shows that convex loss functions such as the LAD-LASSO or penalized M-estimators are not robust enough and even a single outlier can send the regression coefficients to infinity (Alfons et al, 2013). On the other hand, bounded nonconvex loss functions have shown more resistant to outliers than traditional convex loss functions in regression and classification (Liu et al, 2007;Freund, 2001;Wu and Liu, 2007;Singh et al, 2014;Chi and Scott, 2014;Li and Bradic, 2018;Wang, 2018a,b). This article is concerned with variable selection using penalized nonconvex loss functions in regression and classification.…”
Section: Introductionmentioning
confidence: 99%