2022
DOI: 10.48550/arxiv.2203.00953
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Computationally Efficient and Statistically Optimal Robust Low-rank Matrix and Tensor Estimation

Abstract: Low-rank matrix estimation under heavy-tailed noise is challenging, both computationally and statistically. Convex approaches have been proven statistically optimal but suffer from high computational costs, especially since robust loss functions are usually non-smooth. More recently, computationally fast non-convex approaches via sub-gradient descent are proposed, which, unfortunately, fail to deliver a statistically consistent estimator even under sub-Gaussian noise. In this paper, we introduce a novel Rieman… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 16 publications
0
5
0
Order By: Relevance
“…The Huber loss has been widely used in various regression models such as Fan et al (2017) [3] and Sun et al (2020) [8] for linear regression, Tan et al (2022) [9] for reduced rank regression and Shen et al (2022) [7] for matrix recovery.…”
Section: Resultsmentioning
confidence: 99%
“…The Huber loss has been widely used in various regression models such as Fan et al (2017) [3] and Sun et al (2020) [8] for linear regression, Tan et al (2022) [9] for reduced rank regression and Shen et al (2022) [7] for matrix recovery.…”
Section: Resultsmentioning
confidence: 99%
“…It is unlikely that we can remove this additional term. Indeed, similar phenomenon exists ubiquitously non-convex algorithms in the literature, such as low rank matrix recovery [44,52] or low rank tensor recovery [12,41,45]. For example, in the matrix case, we can analogously define the stable rank of a rank-r matrix M ∈ R n×n as sr(M)…”
Section: Discussion On Sample Complexitymentioning
confidence: 96%
“…where in the last inequality we used (41). The above inequality also implies dist(x 0 , w 0 t ) ⩽ 1.…”
Section: Proof Of Theoremmentioning
confidence: 95%
“…To enhance robustness, we truncate overlarge responses to robustify the quadratic loss. Note that compared with previous works [12,18,21,46,53,59] that revolve around highdimensional structured signal recovery and convex programming, our problem setting is new and significantly different. To be specific, the objective is non-convex and x 0 is unstructured.…”
Section: Ngpr With Heavy-tailed Random Noisementioning
confidence: 99%
“…To address the issue, many techniques in statistical estimation have been developed, especially in mean estimation, including median of means [25,43] and Catoni's mean estimator via robust empirical loss [4,11], also see the survey [38]. Particularly, a shrinkage principle for low-rank trace regression was recently developed in [18], which was then applied in corrupted generalized linear regression [59], matrix and tensor completion [46], one-bit estimation [12], vector autoregressive model [53].…”
Section: Ngpr With Heavy-tailed Random Noisementioning
confidence: 99%