2016
DOI: 10.1177/1471082x16642560
|View full text |Cite
|
Sign up to set email alerts
|

Regularized regression for categorical data

Abstract: In the last two decades, regularization techniques, in particular penalty-based methods, have become very popular in statistical modelling. Driven by technological developments, most approaches have been designed for high-dimensional problems with metric variables, whereas categorical data has largely been neglected. In recent years, however, it has become clear that regularization is also very promising when modelling categorical data. A specific trait of categorical data is that many parameters are typically… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
50
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 60 publications
(54 citation statements)
references
References 116 publications
0
50
0
Order By: Relevance
“…On the other hand, regression obtained the best results for the training data set, which suggests that it may be overfitting the training data. As pointed out by an anonymous referee, a penalty item is commonly inserted into the error function as a regularization method when fitting a regression model to data, in order to prevent overfitting [40][41][42]. We do not added such a penalty in our analysis so that a fairer comparison between the methods can be made given that regularization methods have not been proposed yet in the empirical similarity methodology.…”
Section: Discussionmentioning
confidence: 99%
“…On the other hand, regression obtained the best results for the training data set, which suggests that it may be overfitting the training data. As pointed out by an anonymous referee, a penalty item is commonly inserted into the error function as a regularization method when fitting a regression model to data, in order to prevent overfitting [40][41][42]. We do not added such a penalty in our analysis so that a fairer comparison between the methods can be made given that regularization methods have not been proposed yet in the empirical similarity methodology.…”
Section: Discussionmentioning
confidence: 99%
“…In addition, given the presence of a potentially large number of categorical regressors in the context of road safety studies, it would be beneficial to investigate possible approaches to variable selection. A promising avenue could be the incorporation in our framework of further regularizations as explained, for example, by Tutz and Gertheiss (2016).…”
Section: Discussionmentioning
confidence: 99%
“…Notwithstanding those results, Q λ could be potentially extended also to incorporate further regularizations. Recently Tutz and Gertheiss (2016) reviewed the use of alternative penalizations (e.g. the fused and group lasso) to perform enhanced estimation and variable selection in models with categorical responses and predictors.…”
Section: Penalized Generalized Linear Model Representationmentioning
confidence: 99%
“…Specification of a Bayesian linear regression model requires not only a model for the data, for example, the linear regression model (3) of Tutz and Gertheiss (2016), y =˛+ p j=1 X jˇj + ε, ε ∼ N(0, 2 I),…”
Section: Bayesian Regularization Of Effects Of Categorical Covariatesmentioning
confidence: 99%