2018
DOI: 10.1007/s10589-018-0034-y
|View full text |Cite
|
Sign up to set email alerts
|

Structured nonconvex and nonsmooth optimization: algorithms and iteration complexity analysis

Abstract: Nonconvex and nonsmooth optimization problems are frequently encountered in much of statistics, business, science and engineering, but they are not yet widely recognized as a technology in the sense of scalability. A reason for this relatively low degree of popularity is the lack of a well developed system of theory and algorithms to support the applications, as is the case for its convex counterpart. This paper aims to take one step in the direction of disciplined nonconvex and nonsmooth optimization. In part… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
152
0
1

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 124 publications
(155 citation statements)
references
References 55 publications
2
152
0
1
Order By: Relevance
“…Problem ǫ-optimality measure Convergence rate Algorithms for Convex Models PGM [16] (4) objective value error O(1/k) APGM [16] (4) objective value error O(1/k 2 ) IALM [17] (2) -convergence, no rate given ADMM [18] (2) -convergence, no rate given ALM [19] (17) objective value error O(1/k) FALM [19] (17) objective value error O(1/k 2 ) ASALM [20] (19) -convergence unclear, no rate given VASALM [20] (19) -convergence, no rate given PSPG [21] (3) objective value error O(1/k) ADMIP [22] (3) objective value error O(1/k) Quasi-Newton method (fastRPCA) [23] (26) -convergence, no rate given 3-block ADMM [24] (28) -convergence, no rate given Frank-Wolfe [25] (30) objective value error O(1/k) Algorithms for Nonconvex Models GoDec [26] (33) -local convergence, no rate given GreBsmo [27] (36) -convergence unclear, no rate given Alternating Minimization (R2PCP) [28] (35) -local convergence, no rate given Gradient Descent (GD) [11] ≈ (36) -linear convergence with proper initialization and incoherence assumption Alternating Minimization [29] (37) -local convergence with proper initialization and incoherence and RIP assumptions Stochastic alg. [30] (39) -convergence if the iterates are always full rank matrices, no rate given LMafit [31] (44) -convergence if difference between two consecutive iterates tends to zero, no rate given Conditional Gradient [32] (48) [32] (54) perturbed KKT O(1/ √ k) † Note: Some of these algorithms solve different problems and the ǫ-optimality measures are also different, so the convergence rates are not directly comparable with each other. [11] has no explicit optimization formulation but the objective is similar to (36).…”
Section: Algorithmmentioning
confidence: 99%
“…Problem ǫ-optimality measure Convergence rate Algorithms for Convex Models PGM [16] (4) objective value error O(1/k) APGM [16] (4) objective value error O(1/k 2 ) IALM [17] (2) -convergence, no rate given ADMM [18] (2) -convergence, no rate given ALM [19] (17) objective value error O(1/k) FALM [19] (17) objective value error O(1/k 2 ) ASALM [20] (19) -convergence unclear, no rate given VASALM [20] (19) -convergence, no rate given PSPG [21] (3) objective value error O(1/k) ADMIP [22] (3) objective value error O(1/k) Quasi-Newton method (fastRPCA) [23] (26) -convergence, no rate given 3-block ADMM [24] (28) -convergence, no rate given Frank-Wolfe [25] (30) objective value error O(1/k) Algorithms for Nonconvex Models GoDec [26] (33) -local convergence, no rate given GreBsmo [27] (36) -convergence unclear, no rate given Alternating Minimization (R2PCP) [28] (35) -local convergence, no rate given Gradient Descent (GD) [11] ≈ (36) -linear convergence with proper initialization and incoherence assumption Alternating Minimization [29] (37) -local convergence with proper initialization and incoherence and RIP assumptions Stochastic alg. [30] (39) -convergence if the iterates are always full rank matrices, no rate given LMafit [31] (44) -convergence if difference between two consecutive iterates tends to zero, no rate given Conditional Gradient [32] (48) [32] (54) perturbed KKT O(1/ √ k) † Note: Some of these algorithms solve different problems and the ǫ-optimality measures are also different, so the convergence rates are not directly comparable with each other. [11] has no explicit optimization formulation but the objective is similar to (36).…”
Section: Algorithmmentioning
confidence: 99%
“…Apart from previous work discussed above, many researches concerned with optimization problems with a general class of nonconvex regularizations [12], [13], [19], [20] are developed. Nevertheless, these proposed methods cannot be applied to solve the optimization problem studied in this paper.…”
Section: Related Workmentioning
confidence: 99%
“…Gong et al [12] proposed General Iterative Shrinkage and Thresholding (GIST) algorithm to solve the nonconvex optimization problem for a large class of nonconvex penalties. Recently, Jiang et al [13] have proposed two proximal-type variants of the ADMM to solve the structured nonconvex and nonsmooth problems. Nevertheless, the algorithms proposed in [12] and [13] are unable to solve the nonconvex penalized hinge loss function because they both require the loss function to be differentiable as well.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations