2020
DOI: 10.1109/tcyb.2018.2874332
|View full text |Cite
|
Sign up to set email alerts
|

Primal Averaging: A New Gradient Evaluation Step to Attain the Optimal Individual Convergence

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
29
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1
1

Relationship

2
5

Authors

Journals

citations
Cited by 20 publications
(29 citation statements)
references
References 20 publications
0
29
0
Order By: Relevance
“…It is easy to find that the gradient operation in (11) is imposed on f (w t ), in which w t in fact is a weighted average of all past iterates. When the objective in (1) is μ-strongly convex, stochastic PA-PSG [29] is modified as ⎧ ⎪ ⎪ ⎨ ⎪ ⎪ ⎩…”
Section: Psg In Nonsmooth Optimizationmentioning
confidence: 99%
See 1 more Smart Citation
“…It is easy to find that the gradient operation in (11) is imposed on f (w t ), in which w t in fact is a weighted average of all past iterates. When the objective in (1) is μ-strongly convex, stochastic PA-PSG [29] is modified as ⎧ ⎪ ⎪ ⎨ ⎪ ⎪ ⎩…”
Section: Psg In Nonsmooth Optimizationmentioning
confidence: 99%
“…In [29], PA-like algorithms are extended to solve regularized nonsmooth loss optimization problems in stochastic settings. Specifically, PA-PSG (11) is reformulated as ⎧ ⎨ ⎩…”
Section: Extension To Regularized Learningmentioning
confidence: 99%
“…In [29], PA-like algorithms are extended to solve regularized nonsmooth loss optimization problems in stochastic settings. Specifically, PA-PSG (11) is reformulated as ⎧ ⎨ ⎩ w + t = arg min w∈Q a t ĝ t , w +γ t B w, w + t −1 +a t r (w)…”
Section: Extension To Regularized Learningmentioning
confidence: 99%
“…It should be mentioned that the derivations of optimal individual convergence so far only focus on the DA-like methods. Motivated by the averaging step in quasi-monotone DA [22], we recently presented a primal averaging (PA) strategy for PSG [29], in which the subgradient evaluation is imposed on the average of all past iterates. The PA strategy can accelerate the individual convergence of PSG to be optimal for nonsmooth convex problems.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation