2017
DOI: 10.1287/moor.2016.0817
|View full text |Cite
|
Sign up to set email alerts
|

A Descent Lemma Beyond Lipschitz Gradient Continuity: First-Order Methods Revisited and Applications

Abstract: The proximal gradient and its variants is one of the most attractive first-order algorithm for minimizing the sum of two convex functions, with one being nonsmooth. However, it requires the differentiable part of the objective to have a Lipschitz continuous gradient, thus precluding its use in many applications. In this paper we introduce a framework which allows to circumvent the intricate question of Lipschitz continuity of gradients by using an elegant and easy to check convexity condition which captures th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
406
0
1

Year Published

2017
2017
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 265 publications
(409 citation statements)
references
References 34 publications
2
406
0
1
Order By: Relevance
“…For relatively prox-regular functions, our results guarantee stationarity with respect to the equivalent inf-projected problem that involves the Bregman-Moreau envelope. Based on the gradient formulas for the Bregman-Moreau envelope, we conclude this section by highlighting a local equivalence between alternating minimization and recent Bregman proximal gradient algorithms [6,18,5].…”
Section: Outline and Summary Of Our Contributionmentioning
confidence: 98%
See 2 more Smart Citations
“…For relatively prox-regular functions, our results guarantee stationarity with respect to the equivalent inf-projected problem that involves the Bregman-Moreau envelope. Based on the gradient formulas for the Bregman-Moreau envelope, we conclude this section by highlighting a local equivalence between alternating minimization and recent Bregman proximal gradient algorithms [6,18,5].…”
Section: Outline and Summary Of Our Contributionmentioning
confidence: 98%
“…Moreover, in analogy to the Euclidean setting, for which strongly amenable functions provide a source of examples for prox-regular functions, we introduce relatively amenable functions. Their definition is based on a recent generalization of functions that have a Lipschitz continuous gradient to L-smooth adaptable (or relatively smooth) functions [6,18], i.e., functions that are convex relative to a function that generates the Bregman distance.…”
Section: Motivation and Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…This is not true for the Poisson Log-Likelihood. However, (Bauschke et al, 2016) utilized a different definition of smoothness and derived a generalized algorithm (NoLips) that is capable of achieving a sublinear rate of convergence for a class of non-Lipschitz continuous functions including the Poisson Log-Likelihood. Here we use a custom alternating minimization approach using the NoLips algorithm to optimize the Poisson Log-Likelihood with additional modifications to allow for faster computation for sparse matrices and ability to parallelize the computation (see supplementary materials).…”
Section: State Estimationmentioning
confidence: 99%
“…UNCURL is capable of running on larger datasets comprising of up to millions of cells. UNCURL uses a fast public NMF package (Pedregosa et al, 2011) for the log-normal and Gaussian distributions, while using SPNoLips (Sparse-Parallel-NoLips, described in supplementary methods), which is a custom implementation based on the NoLips algorithm (Bauschke et al (2016)) for the Poisson distribution. Both implementations are capable of using sparse matrices as input for memory and runtime advantages, and are parallelizable.…”
Section: Scalabilitymentioning
confidence: 99%