2019
DOI: 10.1080/10556788.2019.1637433
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive online distributed optimization in dynamic environments

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
3

Relationship

3
6

Authors

Journals

citations
Cited by 13 publications
(9 citation statements)
references
References 26 publications
0
9
0
Order By: Relevance
“…For example, the momentum acceleration technique was exploited to improve the performance under time-varying unbalanced communication graphs in [128], where an improved static regret bound O( √ 1 + log T + √ T ) is established for convex cost functions. Also, adaptive gradient methods have been integrated into DOL in [129]- [131].…”
Section: Metricmentioning
confidence: 99%
See 1 more Smart Citation
“…For example, the momentum acceleration technique was exploited to improve the performance under time-varying unbalanced communication graphs in [128], where an improved static regret bound O( √ 1 + log T + √ T ) is established for convex cost functions. Also, adaptive gradient methods have been integrated into DOL in [129]- [131].…”
Section: Metricmentioning
confidence: 99%
“…The most popular adaptive gradient method is Adam by estimating first-and second-order moments of gradients [139]. As a result, one natural idea is to apply adaptive methods to DOL for improving the performance, which is exactly done in [129]- [131]. However, the research along this line is not far fully explored, leaving an enormous possibility for future directions.…”
Section: Future Directionsmentioning
confidence: 99%
“…Comparing with the SGD method, ADAGRAD dynamically interpolates knowledge of history gradients to adaptively change the learning rate, thereby achieving significantly better performance when the gradients are sparse, or in general small. Another state-of-the-art algorithm for training deep learning models is known as ADAM [18,29], which is a variant of the general class of ADAGRAD-type algorithms. When 0 < β 2 < 1, 0 ≤…”
Section: Adagrad and Adam For Minimization Problemsmentioning
confidence: 99%
“…To overcome this issue, several improved variants of SGD that automatically update the search directions and learning rates using a metric constructed from the history of iterates have been proposed. The pioneering work in this line of research is [23], while recent developments include [12,47,43,24,11,36,33,32,31].…”
Section: Introductionmentioning
confidence: 99%