Our system is currently under heavy load due to increased usage. We're actively working on upgrades to improve performance. Thank you for your patience.
2016
DOI: 10.1007/s10107-016-1017-3
|View full text |Cite
|
Sign up to set email alerts
|

Stochastic compositional gradient descent: algorithms for minimizing compositions of expected-value functions

Abstract: Classical stochastic gradient methods are well suited for minimizing expected-value objective functions. However, they do not apply to the minimization of a nonlinear function involving expected values or a composition of two expected-value functions, i.e., problems of the form min. In order to solve this stochastic composition problem, we propose a class of stochastic compositional gradient descent (SCGD) algorithms that can be viewed as stochastic versions of quasi-gradient method. SCGD update the solutions … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

6
343
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
5
3

Relationship

2
6

Authors

Journals

citations
Cited by 175 publications
(349 citation statements)
references
References 29 publications
6
343
0
Order By: Relevance
“…For the case where T = 1, our results match the best known sample complexity upper-and lower-bounds. For the case where T = 2, our results improve the convergence rate from O(n 2/9 ) of the a-SCGD in [26] to O(n 2/5 ). Besides, with additional assumption that the inner level function f (T ) in (1.1) has Lipschitz continuous gradients, we obtain a convergence rate O(n 4/9 ) for two-level problems, which matches the stateof-art result achieved by ASC-PG in [28].…”
mentioning
confidence: 62%
See 1 more Smart Citation
“…For the case where T = 1, our results match the best known sample complexity upper-and lower-bounds. For the case where T = 2, our results improve the convergence rate from O(n 2/9 ) of the a-SCGD in [26] to O(n 2/5 ). Besides, with additional assumption that the inner level function f (T ) in (1.1) has Lipschitz continuous gradients, we obtain a convergence rate O(n 4/9 ) for two-level problems, which matches the stateof-art result achieved by ASC-PG in [28].…”
mentioning
confidence: 62%
“…We have also obtained convergence and rate of convergence results for nonconvex problems. Table 1 summarizes our results and compare them with the best known ones for the single-and two-level stochastic compositional optimization problems [9,19,23,26,28]. We also provide numerical experiments with a risk-averse regression problem.…”
mentioning
confidence: 98%
“…(1.5) and (1.6)) in optimization, see [27]. Stochastic compositional problems have also appeared in the parallel line of work [66]. There, the authors require the entire composite function to be either convex or smooth.…”
mentioning
confidence: 99%
“…An established approach was to use two-level stochastic recursive algorithms with two stepsize sequences in different time scales: a slower one for updating the main decision variable x, and a faster one for tracking the value of the inner function(s). References [38,39] provide a detailed account of these techniques and existing results. In [40] these ideas were extended to multilevel problems of form (1), albeit with multiple time scales and under continuous differentiability assumptions.…”
Section: Introductionmentioning
confidence: 99%
“…Denote d k = y k − x k . By a discrete-time version of the argument that lead to (38), the decrease of the first part of the Lyapunov function can be estimated as follows:…”
mentioning
confidence: 99%