2018
DOI: 10.1007/s10107-018-1236-x
|View full text |Cite
|
Sign up to set email alerts
|

Minimization of transformed $$L_1$$ L 1 penalty: theory, difference of convex function algorithm, and robust application in compressed sensing

Abstract: We study the minimization problem of a non-convex sparsity promoting penalty function, the transformed l 1 (TL1), and its application in compressed sensing (CS). The TL1 penalty interpolates l 0 and l 1 norms through a nonnegative parameter a ∈ (0, +∞), similar to lp with p ∈ (0, 1], and is known to satisfy unbiasedness, sparsity and Lipschitz continuity properties. We first consider the constrained minimization problem, and discuss the exact recovery of l 0 norm minimal solution based on the null space proper… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
77
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3

Relationship

2
5

Authors

Journals

citations
Cited by 106 publications
(78 citation statements)
references
References 47 publications
0
77
0
Order By: Relevance
“…The inequality (1.6) implies that L 1 may not perform well for highly coherent matrices, i.e., µ(A) ∼ 1, as x 0 is then at most one, which seldom occurs simultaneously with Ax * = b. Other than the popular L 1 norm, there are a variety of regularization functionals to promote sparsity, such as L p [9,43,23], L 1 -L 2 [44,26], capped L 1 (CL1) [48,37], and transformed L 1 (TL1) [29,46,47]. Most of these models are nonconvex, leading to difficulties in proving exact recovery guarantees and algorithmic convergence, but they tend to give better empirical results compared to the convex L 1 approach.…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…The inequality (1.6) implies that L 1 may not perform well for highly coherent matrices, i.e., µ(A) ∼ 1, as x 0 is then at most one, which seldom occurs simultaneously with Ax * = b. Other than the popular L 1 norm, there are a variety of regularization functionals to promote sparsity, such as L p [9,43,23], L 1 -L 2 [44,26], capped L 1 (CL1) [48,37], and transformed L 1 (TL1) [29,46,47]. Most of these models are nonconvex, leading to difficulties in proving exact recovery guarantees and algorithmic convergence, but they tend to give better empirical results compared to the convex L 1 approach.…”
mentioning
confidence: 99%
“…For example, it was reported in [44,26] that L p gives superior results for incoherent matrices (i.e., µ(A) is small), while L 1 -L 2 is the best for the coherent scenario. In addition, TL1 is always the second best no matter whether the matrix is coherent or not [46,47].…”
mentioning
confidence: 99%
“…Obviously, when a → 0, i min(|x i |, a)/a → x 0 . Transformed 1 , which is a smooth version of capped 1 , is discussed in the works [39][40][41]. Some other non-convex metrics with concise form are also considered as alternatives to improve 1 , including p with p ∈ (0, 1) [33][34][35], whose formula is…”
Section: Non-convex Regularization Functionmentioning
confidence: 99%
“…Although 1 enjoys several good properties, it is sensitive to outliers and may cause serious bias in estimation [26,27]. To overcome this defect, many non-convex surrogates are proposed and analyzed, including smoothly clipped absolute deviation (SCAD) [26], log penalty [28,29], capped 1 [30,31], minimax concave penalty (MCP) [32], p penalty with p ∈ (0, 1) [33][34][35], the difference of 1 and 2 norms [36][37][38] and transformed 1 [39][40][41]. More and more works have shown the good performance of non-convex regularizers in both theoretical analyses and practical applications.…”
mentioning
confidence: 99%
“…Since lim a→0 + ρ a (x i ) = 1 {xi =0} , lim a→+∞ ρ a (x i ) = |x i |, ∀i, the T 1 penalty interpolates 1 and 0 . For its sparsification in compressed sensing and other applications, see [14] and references therein. To sparsify weights in GSRNN training via 1 and T 1 , we add them to the loss function of GSRNN with a multiplicative penalty parameter α > 0, and call stochastic gradient descent optimizer on Tensorflow.…”
Section: Sparsity Promoting Penaltiesmentioning
confidence: 99%