2020
DOI: 10.1111/rssb.12359
|View full text |Cite
|
Sign up to set email alerts
|

Right Singular Vector Projection Graphs: Fast High Dimensional Covariance Matrix Estimation under Latent Confounding

Abstract: Summary We consider the problem of estimating a high dimensional p×p covariance matrix Σ, given n observations of confounded data with covariance Σ+ΓnormalΓsans-serifT, where Γ is an unknown p×q matrix of latent factor loadings. We propose a simple and scalable estimator based on the projection onto the right singular vectors of the observed data matrix, which we call right singular vector projection (RSVP). Our theoretical analysis of this method reveals that, in contrast with approaches based on the removal … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
7
2

Relationship

2
7

Authors

Journals

citations
Cited by 16 publications
(14 citation statements)
references
References 47 publications
(123 reference statements)
0
14
0
Order By: Relevance
“…We used the 'muscle-skeletal' dataset from the GTEx project 2 , which has 491 rows and 14 713 columns, as our uncorrupted dataset. We preprocessed the data as in Shah et al [2020] by regressing out measured and estimated confounders. We took as a response variable a column randomly selected (anew in each run) from the matrix, meaning that for our experiment n = 491 and p = 14 712.…”
Section: Corrupted Datamentioning
confidence: 99%
“…We used the 'muscle-skeletal' dataset from the GTEx project 2 , which has 491 rows and 14 713 columns, as our uncorrupted dataset. We preprocessed the data as in Shah et al [2020] by regressing out measured and estimated confounders. We took as a response variable a column randomly selected (anew in each run) from the matrix, meaning that for our experiment n = 491 and p = 14 712.…”
Section: Corrupted Datamentioning
confidence: 99%
“…The variables A would encode the first few principal components of X, and hence A would not be exogenous: then OLS+A (γ = 0) is adjusting for these first principal components and aims to remove hidden confounding bias in estimating the causal parameter β. This common practice in applied statistics (e.g., Novembre et al, 2008) can be justified under the assumption of "dense confounding" where the hidden variables H affect most of the X components ( Ćevid, Bühlmann and Meinshausen, 2018); see also closely related work by, for example, Chandrasekaran, Parrilo and Willsky (2012), Shah et al (2018), Guo, Ćevid and Bühlmann (2020). The theoretical and methodological arguments are different since A is now a proxy for the hidden latent confounder H , very different from a valid instrument and not exogenous.…”
Section: Choosing the Amount Of Causal Regularization Amplification Bias And Specification Of Anchorsmentioning
confidence: 99%
“…We mention here that Shah et al (2020) provide vaguely related results on robustness for the GTEx data for another Ridge-type procedure for undirected graphical models.…”
Section: An Illustration On Data From the Gtex Consortiummentioning
confidence: 99%