2013
DOI: 10.1007/s10107-013-0725-1
|View full text |Cite
|
Sign up to set email alerts
|

Dual subgradient algorithms for large-scale nonsmooth learning problems

Abstract: International audience"Classical" First Order (FO) algorithms of convex optimization, such as Mirror Descent algorithm or Nesterov's optimal algorithm of smooth convex optimization, are well known to have optimal (theoretical) complexity estimates which do not depend on the problem dimension. However, to attain the optimality, the domain of the problem should admit a "good proximal setup". The latter essentially means that (1) the problem domain should satisfy certain geometric conditions of "favorable geometr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
40
0

Year Published

2014
2014
2022
2022

Publication Types

Select...
7
1

Relationship

2
6

Authors

Journals

citations
Cited by 22 publications
(40 citation statements)
references
References 22 publications
0
40
0
Order By: Relevance
“…and therefore the spectral filtering proposed above (Eqs. (16), (19), (20) ,which hold for the general one-homogeneous case) decomposes f correctly.…”
Section: Decompositionmentioning
confidence: 98%
See 2 more Smart Citations
“…and therefore the spectral filtering proposed above (Eqs. (16), (19), (20) ,which hold for the general one-homogeneous case) decomposes f correctly.…”
Section: Decompositionmentioning
confidence: 98%
“…1 an example of spectral TV processing is shown with the response of the four filters defined above in Eqs. (19) through (22).…”
Section: The Tv Transformmentioning
confidence: 99%
See 1 more Smart Citation
“…First introduced by Nemirovski and Yudin [39] for non-smooth problems, the mirror descent algorithm and its variants have met with prolific success in convex programming [9], online and stochastic optimization [46], variational inequalities [40], non-cooperative games [19,38], and many other fields of optimization theory and its applications. Nevertheless, despite the appealing convergence properties of (MD), it is often difficult to calculate the update step from x to x + when the problem's feasible region X is not "prox-friendly" -i.e., when there is no efficient oracle for solving the convex optimization problem in (MD) [24]. With this in mind, our main goal in this paper is to provide a convergent, forward discretization of (HRGD) which does not require solving a convex optimization problem at each update step.…”
Section: Introductionmentioning
confidence: 99%
“…The problems of minimizing continuously differentiable convex functions with Lipschitz continuous gradients.These two classes of convex problems can also be reformulated as structured convex problems, which have been receiving much attention in terms of both theoretical and application aspects. In particular, studies of (sub)gradient-based methods for the class of "smoothable" functions [1,6,9,27,35,36], the class of composite problems [1,5,8,17,18,19,26,38,42,43], and the class of weakly smooth problems [11,12,39,40] are notably important.In this paper, we particularly focus on the following two kinds of (sub)gradient methods: the Proximal (sub)Gradient Method (PGM) and the Conditional Gradient Method (CGM). Both methods may require easy-to-solve subproblems at each iteration.The PGM is executed using a prox-function to define a reasonable proximal operator.…”
mentioning
confidence: 99%