2016
DOI: 10.1109/tit.2016.2555812
|View full text |Cite
|
Sign up to set email alerts
|

Model Selection and Minimax Estimation in Generalized Linear Models

Abstract: We consider model selection in generalized linear models (GLM) for high-dimensional data and propose a wide class of model selection criteria based on penalized maximum likelihood with a complexity penalty on the model size. We derive a general nonasymptotic upper bound for the Kullback-Leibler risk of the resulting estimators and establish the corresponding minimax lower bounds for sparse GLM. For the properly chosen (nonlinear) penalty, the resulting penalized maximum likelihood estimator is shown to be asym… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
6
2
1

Relationship

2
7

Authors

Journals

citations
Cited by 21 publications
(18 citation statements)
references
References 26 publications
0
18
0
Order By: Relevance
“…Although logistic regression is widely used in various classification problems, its rigorous theoretical ground has not been yet properly established. Model selection in a general framework of generalized linear models (GLM) and in logistic regression in particular was studied in Abramovich and Grinshtein (2016). They proposed model selection procedure based on penalized maximum likelihood with a complexity penalty on the model size and investigated the goodness-of-fit of the resulting estimator in terms of the Kullback-Leibler risk.…”
Section: Introductionmentioning
confidence: 99%
“…Although logistic regression is widely used in various classification problems, its rigorous theoretical ground has not been yet properly established. Model selection in a general framework of generalized linear models (GLM) and in logistic regression in particular was studied in Abramovich and Grinshtein (2016). They proposed model selection procedure based on penalized maximum likelihood with a complexity penalty on the model size and investigated the goodness-of-fit of the resulting estimator in terms of the Kullback-Leibler risk.…”
Section: Introductionmentioning
confidence: 99%
“…The generalized logistic regression analysis model also contains a random component with the Bernoulli distribution to characterize the stochastic effects [21]. The logit link function of generalized logistic regression analysis calculates the natural logarithm of an odds ratio of the binomial probabilities, which can be written asT1lnPIPDPCO=xTβ=β0+x1β1++xfβf,where P IPD and P CO = 1 − P IPD denote the probabilities of binary classes (i.e., IPD and CO subject groups), the vector β = [ β 0 , β 1 ,…, β f ] T represents the generalized logistic regression coefficients, and x = [1, x 1 ,…, x f ] T is the model input vector including unity and five selected vocal features, the latter of which include MDVP:F0, Spread1, MDVP-LDA, Shimmer-LDA, and Nonlinear-LDA.…”
Section: Methodsmentioning
confidence: 99%
“…For most models belonging the canonical exponential family, the step III. (ii) is quite trivial, see Lemma 1 in Abramovich and Grinshtein (2016) for example. Nonetheless, it is worthy to note that the our loss function is not in the canonical exponential family, so there is no extent discussion about the lower bound of likelihood-based divergence of θ and θ * in our setting.…”
Section: (Iii) the Last Term In (A2)mentioning
confidence: 99%