2021
DOI: 10.48550/arxiv.2102.09080
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Adjusting the Benjamini-Hochberg method for controlling the false discovery rate in knockoff assisted variable selection

Abstract: This paper revisits the knockoff-based multiple testing setup considered in Barber & Candès (2015) for variable selection applied to a linear regression model with n ≥ 2d, where n is the sample size and d is the number of explanatory variables. The BH method based on ordinary least squares estimates of the regressions coefficients is adjusted to this setup, making it a valid p-value based FDR controlling method that does not rely on any specific correlation structure of the explanatory variables. Simulations a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 7 publications
0
4
0
Order By: Relevance
“…How can practitioners recognize which is better for their context? Can hybrid methods such as those of Sarkar and Tang (2021), or methods yet to be developed, balance the tradeoffs between the two approaches, preserving the strengths of the knockoffs framework without suffering its drawbacks? By identifying pitfalls for the knockoffs framework our results represent strides toward a more complete understanding of multiple testing in the linear model.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…How can practitioners recognize which is better for their context? Can hybrid methods such as those of Sarkar and Tang (2021), or methods yet to be developed, balance the tradeoffs between the two approaches, preserving the strengths of the knockoffs framework without suffering its drawbacks? By identifying pitfalls for the knockoffs framework our results represent strides toward a more complete understanding of multiple testing in the linear model.…”
Section: Discussionmentioning
confidence: 99%
“…To begin to explain how knockoffs can go wrong, Section 1.2 formally recasts the knockoff filter as a conditional post-selection inference method built around a randomized estimator β = β + ω, where ω is user-generated Gaussian noise in the style of Tian and Taylor (2018). Our interpretation builds on a conditioning argument in Barber and Candès (2019) and an observation in Sarkar and Tang (2021) that knockoffs constructs two independent estimators for β.…”
Section: Introductionmentioning
confidence: 99%
“…More recently, randomization was used in Li and Fithian (2021) to recast the knockoff procedure of Barber and Candès (2015) as a selective inference procedure for the linear Gaussian model that adds noise to the OLS estimates ( β) to create a "whitened" version of β to use for hypothesis selection. The work of Sarkar and Tang (2021) explores similar ways of using knockoffs to split β into independent Full Dataset (X) Selection Event S(X)…”
Section: Related Work On Data Splitting and Carvingmentioning
confidence: 99%
“…Second, after the test statistic is determined for each hypothesis, how to combine these test statistics to derive a method that controls the false discovery rate is challenging. Many existing procedures, such as Ji and Zhao (2014); Candès (2015, 2019); Xing et al (2021); Sarkar and Tang (2021), work on the (generalized) linear regression models. In Candes et al (2018), the authors considered an arbitrary joint distribution of y and x and proposed the model-X knockoff to control FDR.…”
Section: Introductionmentioning
confidence: 99%