2022
DOI: 10.1093/imaiai/iaac024
|View full text |Cite
|
Sign up to set email alerts
|

Nearly minimax-optimal rates for noisy sparse phase retrieval via early-stopped mirror descent

Abstract: This paper studies early-stopped mirror descent applied to noisy sparse phase retrieval, which is the problem of recovering a $k$-sparse signal $\textbf{x}^\star \in{\mathbb{R}}^n$ from a set of quadratic Gaussian measurements corrupted by sub-exponential noise. We consider the (non-convex) unregularized empirical risk minimization problem and show that early-stopped mirror descent, when equipped with the hypentropy mirror map and proper initialization, achieves a nearly minimax-optimal rate of convergence, pr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(7 citation statements)
references
References 35 publications
0
7
0
Order By: Relevance
“…Especially, when there are only very few significant components in x such that sx = O(1), the sample complexity of algorithm 1 is m = Ω(s log n), which is optimal in s and order-wisely better than m = Ω(s 2 log(n)) in [25,50]. Recently, [53] proposed a new initialization method that is a multiple run of algorithm 2 combined with a few steps of Hardmard Wirtinger flow, and the sample complexity of its initialization is m = Ω(δ −2 sx s log(n)) under the assumption that min i∈S |x i | ⩾ ∥x∥ ℓ2 / √ s. On the contrary, algorithm 1 not only achieves the same order of sample complexity without any assumption on x, but also improves dependency on δ and sx from δ −2 sx to max{δ −2 ,s x }. Since δ is usually a small constant in various non-convex algorithms, this improvement is significant, which makes algorithm 1 initialized algorithms outperform others.…”
Section: Discussion On Sample Complexitymentioning
confidence: 99%
See 4 more Smart Citations
“…Especially, when there are only very few significant components in x such that sx = O(1), the sample complexity of algorithm 1 is m = Ω(s log n), which is optimal in s and order-wisely better than m = Ω(s 2 log(n)) in [25,50]. Recently, [53] proposed a new initialization method that is a multiple run of algorithm 2 combined with a few steps of Hardmard Wirtinger flow, and the sample complexity of its initialization is m = Ω(δ −2 sx s log(n)) under the assumption that min i∈S |x i | ⩾ ∥x∥ ℓ2 / √ s. On the contrary, algorithm 1 not only achieves the same order of sample complexity without any assumption on x, but also improves dependency on δ and sx from δ −2 sx to max{δ −2 ,s x }. Since δ is usually a small constant in various non-convex algorithms, this improvement is significant, which makes algorithm 1 initialized algorithms outperform others.…”
Section: Discussion On Sample Complexitymentioning
confidence: 99%
“…Consequently, it is possible to break the sample complexity barriers as long as the methods for initialization can be improved. Alternatively, a Hadamard Wirtinger flow (HWF) method has been introduced in [53] as a different strategy to obtain the initial guess, where an implicit regularization is used. The sample complexity of HWF is improved to Ω(s x s log n) under the assumption that x min = Ω( 1 √ s ∥x∥ ℓ2 ), where ∥ • ∥ ℓ2 denotes the ℓ 2 -norm for vectors, sx :=…”
Section: Initialization Methods Sample Complexity M Assumptionsmentioning
confidence: 99%
See 3 more Smart Citations