2016
DOI: 10.1080/10556788.2016.1138222
|View full text |Cite
|
Sign up to set email alerts
|

A second-order method for convex1-regularized optimization with active-set prediction

Abstract: We describe an active-set method for the minimization of an objective function φ that is the sum of a smooth convex function f and an 1 -regularization term. A distinctive feature of the method is the way in which active-set identification and second-order subspace minimization steps are integrated to combine the predictive power of the two approaches. At every iteration, the algorithm selects a candidate set of free and fixed variables, performs an (inexact) subspace phase, and then assesses the quality of th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 24 publications
(17 citation statements)
references
References 31 publications
(64 reference statements)
0
17
0
Order By: Relevance
“…Since U ∈ ℜ n×n is a diagonal matrix, at the first glance, the costs of computing AU A T and the matrixvector multiplication AU A T d for a given vector d ∈ ℜ m are O(m 2 n) and O(mn), respectively. These computational costs are too expensive when the dimensions of A are large and can make the commonly employed approaches such as the Cholesky factorization and the conjugate gradient method inappropriate for solving (26). Fortunately, under the sparse optimization setting, if the sparsity of U is wisely taken into the consideration, one can substantially reduce these unfavorable computational costs to a level such that they are negligible or at least insignificant compared to other costs.…”
Section: An Efficient Implementation Of Ssn For Solving Subproblems (18)mentioning
confidence: 99%
See 1 more Smart Citation
“…Since U ∈ ℜ n×n is a diagonal matrix, at the first glance, the costs of computing AU A T and the matrixvector multiplication AU A T d for a given vector d ∈ ℜ m are O(m 2 n) and O(mn), respectively. These computational costs are too expensive when the dimensions of A are large and can make the commonly employed approaches such as the Cholesky factorization and the conjugate gradient method inappropriate for solving (26). Fortunately, under the sparse optimization setting, if the sparsity of U is wisely taken into the consideration, one can substantially reduce these unfavorable computational costs to a level such that they are negligible or at least insignificant compared to other costs.…”
Section: An Efficient Implementation Of Ssn For Solving Subproblems (18)mentioning
confidence: 99%
“…See Figure 2 for an illustration on the computation of A T J A J . In this case, the total computational costs for solving the Newton linear system (26) are reduced significantly further from O(m 2 (m + r)) to O(r 2 (m + r)). We should emphasize here that this dramatic reduction on the computational costs results from the wise combination of the careful examination of the existing second order sparsity in the Lasso-type problems and some "smart" numerical linear algebra.…”
Section: An Efficient Implementation Of Ssn For Solving Subproblems (18)mentioning
confidence: 99%
“…The active set strategies have also been studied in [65,107]. Specifically, the method in [65] solves a smooth quadratic subproblem determined by the active sets and invokes a corrective cycle that greatly improves the efficiency and robustness of the algorithm. The method is globalized by using a proximal gradient step to check the desired progress.…”
Section: Active Set Methodsmentioning
confidence: 99%
“…A subset of the variables is fixed in the so-called active sets determined by certain mechanisms and the remaining variables are computed from carefully constructed subproblems. Examples include optimization problems with bound constraints or linear constraints in [17,18,53,82,83], ℓ 1 -regularized problem for sparse optimization in [65,107,135] and general nonlinear programs in [19,20]. In quadratic programming, the inequality constraints that have zero values at the optimal solution are called active, and they are replaced by equality constraints in the subproblem [113].…”
Section: Bcdmentioning
confidence: 99%
“…These are just two of many algorithms that incorporate working sets to speed up sparse optimization. For lasso-type problems, many additional studies combine working set (Scheinberg and Tang, 2016;Massias et al, 2017) or active set (Wen et al, 2012;Solntsev et al, 2015;Keskar et al) strategies with standard algorithms. Researchers have also applied working sets to many other sparse problems-see e.g.…”
Section: Relation To Prior Algorithmsmentioning
confidence: 99%