2009
DOI: 10.1198/jasa.2009.0101
|View full text |Cite
|
Sign up to set email alerts
|

Generalized Thresholding of Large Covariance Matrices

Abstract: We propose a new class of generalized thresholding operators that combine thresholding with shrinkage, and study generalized thresholding of the sample covariance matrix in high dimensions. Generalized thresholding of the covariance matrix has good theoretical properties and carries almost no computational burden. We obtain an explicit convergence rate in the operator norm that shows the tradeoff between the sparsity of the true model, dimension, and the sample size, and shows that generalized thresholding is … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

7
458
0
1

Year Published

2012
2012
2018
2018

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 415 publications
(485 citation statements)
references
References 28 publications
7
458
0
1
Order By: Relevance
“…We consider the shrinkage towards identity or equicorrelation estimators proposed by Ledoit and Wolf (2004a,b), setting the shrinkage parameters to their respective optimal values derived in the aforementioned papers. We also consider the hard and soft thresholding estimators due to Bickel and Levina (2008b) and Rothman, Levina, and Zhu (2009). The various fine-tuning parameters required to calculate these estimators are set in accordance with the default values of the R code kindly shared with us by Adam J. Rothman.…”
Section: Comparison With Leading Existing Estimatorsmentioning
confidence: 99%
“…We consider the shrinkage towards identity or equicorrelation estimators proposed by Ledoit and Wolf (2004a,b), setting the shrinkage parameters to their respective optimal values derived in the aforementioned papers. We also consider the hard and soft thresholding estimators due to Bickel and Levina (2008b) and Rothman, Levina, and Zhu (2009). The various fine-tuning parameters required to calculate these estimators are set in accordance with the default values of the R code kindly shared with us by Adam J. Rothman.…”
Section: Comparison With Leading Existing Estimatorsmentioning
confidence: 99%
“…The first dataset has 64 training tissue samples with four types of tumors (23 EWS, 8 BL-NHL, 12 NB, and 21 RMS) and 6567 gene expression values for each sample. We applied the prefiltering step used in Khan et al (2001) and then picked the top 40 and bottom 160 genes based on the F-statistic as done in Rothman, Levina, and Zhu (2009). The second dataset has 63 subjects with 44 healthy controls and 19 cardiovascular patients, and 20,426 genes measured for each subject.…”
Section: Comparing Two Positive-definite Thresholding Estimatorsmentioning
confidence: 99%
“…The usual sample covariance matrix is optimal in the classical setting with large samples and fixed low dimensions (Anderson 1984), but it performs very poorly in the highdimensional setting (Marčenko and Pastur 1967;Johnstone 2001). In the recent literature, regularization techniques have been used to improve the sample covariance matrix estimator, including banding (Wu and Pourahmadi 2003;Bickel and Levina 2008a), tapering (Furrer and Bengtsson 2007;Cai, Zhang, and Zhou 2010), and thresholding (Bickel and Levina 2008b;El Karoui 2008;Rothman, Levina, and Zhu 2009). Banding or tapering is very useful when the variables have a natural ordering and off-diagonal entries of the target covariance matrix decay to zero as they move away from the diagonal.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations