2018
DOI: 10.1007/s10463-018-0649-x
|View full text |Cite
|
Sign up to set email alerts
|

AIC for the non-concave penalized likelihood method

Abstract: Non-concave penalized maximum likelihood methods, such as the Bridge, the SCAD, and the MCP, are widely used because they not only perform the parameter estimation and variable selection simultaneously but also are more efficient than the Lasso. They include a tuning parameter which controls a penalty level, and several information criteria have been developed for selecting it. While these criteria assure the model selection consistency, they have a problem in that there are no appropriate rules for choosing o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
13
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(13 citation statements)
references
References 31 publications
0
13
0
Order By: Relevance
“…We may also apply our moment-convergence results to validate the celebrated AIC methodology ( [3], and also [10], [26], and [27]) even under the sparse asymptotics. Recent studies exactly in this direction contain [25] and [29], where the uniform integrability of the sparse maximum-likelihood estimator with the bridge-like regularization played an important role for validating the asymptotic bias correction. Below we will briefly discuss how we can extend the result of [29] to cover a broader range of statistical models with locally asymptotically normal structure.…”
Section: Prediction-related Issuesmentioning
confidence: 99%
See 2 more Smart Citations
“…We may also apply our moment-convergence results to validate the celebrated AIC methodology ( [3], and also [10], [26], and [27]) even under the sparse asymptotics. Recent studies exactly in this direction contain [25] and [29], where the uniform integrability of the sparse maximum-likelihood estimator with the bridge-like regularization played an important role for validating the asymptotic bias correction. Below we will briefly discuss how we can extend the result of [29] to cover a broader range of statistical models with locally asymptotically normal structure.…”
Section: Prediction-related Issuesmentioning
confidence: 99%
“…Recent studies exactly in this direction contain [25] and [29], where the uniform integrability of the sparse maximum-likelihood estimator with the bridge-like regularization played an important role for validating the asymptotic bias correction. Below we will briefly discuss how we can extend the result of [29] to cover a broader range of statistical models with locally asymptotically normal structure. To keep things simple, we are only concerned here with correctly specified models, keeping the multi-scaling setup with θ = (α, β); again we note that cases of single scaling can be handled as well without any essential change, just by ignoring the β-part.…”
Section: Prediction-related Issuesmentioning
confidence: 99%
See 1 more Smart Citation
“…2 An important conclusion of these studies is that sign-constraints themselves can be as effective as more explicit regularization methods such as the Lasso if the design matrix satisfies a special condition called the positive eigenvalue condition (which is called the self-regularizing property in [18]) in addition to a compatibility condition, where the latter condition is a standard one to prove the theoretical properties of regularization methods. More precisely, Meinshausen [14] has derived oracle inequalities for the ℓ 1 -error of the regression 1 To avoid such a computational burden, some authors have developed analytical methods for choosing the tuning parameter λ such as using an information criterion; see [15,20] and references therein for details. 2 We remark that the statistical property of the NNLS estimator in a fixed-dimensional setting has been extensively studied in the literature from a long ago; see e.g.…”
Section: Introductionmentioning
confidence: 99%
“…The importance of such precise estimates of tail probability is well recognized in asymptotic decision theory, prediction, theory of information criteria for model selection, asymptotic expansion, etc. The QLA is rapidly expanding the range of its applications: for example, sampled ergodic diffusion processes (Yoshida [30]), contrastbased information criterion for diffusion processes (Uchida [23]), approximate self-weighted LAD estimation of discretely observed ergodic Ornstein-Uhlenbeck processes (Masuda [12]), jump diffusion processes Ogihara and Yoshida([17]), adaptive estimation for diffusion processes (Uchida and Yoshida [24]), adaptive Bayes type estimators for ergodic diffusion processes (Uchida and Yoshida [27]), asymptotic properties of the QLA estimators for volatility in regular sampling of finite time horizon (Uchida and Yoshida [25]) and in non-synchronous sampling (Ogihara and Yoshida [18]), Gaussian quasi-likelihood random fields for ergodic Lévy driven SDE (Masuda [15]), hybrid multi-step estimators (Kamatani and Uchida [6]), parametric estimation of Lévy processes (Masuda [13]), ergodic point processes for limit order book (Clinet and Yoshida [1]), a non-ergodic point process regression model (Ogihara and Yoshida [19]), threshold estimation for stochastic processes with small noise (Shimizu [21]), AIC for non-concave penalized likelihood method (Umezu et al [28]), Schwarz type model comparison for LAQ models (Eguchi and Masuda [2]), adaptive Bayes estimators and hybrid estimators for small diffusion processes based on sampled data (Nomura and Uchida [16]), moment convergence of regularized leastsquares estimator for linear regression model (Shimizu [22]), moment convergence in regularized estimation under multiple and mixed-rates asymptotics (Masuda and Shimizu [14]), asymptotic expansion in quasi likelihood analysis for volatility (Yoshida [31]) among others.…”
Section: Introductionmentioning
confidence: 99%