The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2019
DOI: 10.1007/s40745-019-00209-4
|View full text |Cite
|
Sign up to set email alerts
|

On Regularisation Methods for Analysis of High Dimensional Data

Abstract: High dimensional data are rapidly growing in many domains due to the development of technological advances which helps collect data with a large number of variables to better understand a given phenomenon of interest. Particular examples appear in genomics, fMRI data analysis, large-scale healthcare analytics, text/image analysis and astronomy. In the last two decades regularisation approaches have become the methods of choice for analysing such high dimensional data. This paper aims to study the performance o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
38
0
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 46 publications
(40 citation statements)
references
References 36 publications
0
38
0
1
Order By: Relevance
“…Identification of a best subset of variables is known to be problematic when the number of explanatory variables is large with respect to the number of subjects and when multicollinearity is present within the data 1 . In this situation, despite their widespread use, it is recognised that selection methods based on exploratory or stepwise procedures using P-values or likelihood-based methods have notable deficiencies including producing inflated coefficient estimates and downward biased errors [1][2][3][4] . This generally results in models that are over fit and with a relatively high number of variables remaining in a 'final' model rather than a sparse model that contains only variables with the greatest association with the outcome 1 .…”
mentioning
confidence: 99%
“…Identification of a best subset of variables is known to be problematic when the number of explanatory variables is large with respect to the number of subjects and when multicollinearity is present within the data 1 . In this situation, despite their widespread use, it is recognised that selection methods based on exploratory or stepwise procedures using P-values or likelihood-based methods have notable deficiencies including producing inflated coefficient estimates and downward biased errors [1][2][3][4] . This generally results in models that are over fit and with a relatively high number of variables remaining in a 'final' model rather than a sparse model that contains only variables with the greatest association with the outcome 1 .…”
mentioning
confidence: 99%
“…The main objective of sparse PCA is to force a number of less important loadings to be zero, resulting in sparse eigenvectors. In order to achieve such sparsity on the extracted components, most of the available methods find the PC's of the covariance matrix through adding a constraint or penalty term from the PCA formulation (1). A constrained l 0 -norm minimisation problem is usually considered as the basic sparse PCA problem as follows: (see also [5])…”
Section: Formulation Of Sparse Pcamentioning
confidence: 99%
“…High dimensional data are rapidly growing in many different disciplines due to the development of technological advances [1]. High dimensional data are particularly common is natural language processing (NLP).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Whereas lasso regression can shrink unnecessary regressors to zero and thereby reduce the number of predictors, ridge regression retains all regressors for inclusion in the model. Both lasso and ridge regression techniques have been shown to perform well when dealing with high-dimensional data under various conditions 64 . Combining the two approaches, elastic-net regression allows for adjustment of the lasso-to-ridge ratio (α), providing greater opportunity for better model fits 65 .…”
Section: Elastic-net Regression Combines Penalty Features Of Lasso Anmentioning
confidence: 99%