2019
DOI: 10.48550/arxiv.1912.01698
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Online and Bandit Algorithms for Nonstationary Stochastic Saddle-Point Optimization

Abhishek Roy,
Yifang Chen,
Krishnakumar Balasubramanian
et al.

Abstract: Saddle-point optimization problems are an important class of optimization problems with applications to game theory, multi-agent reinforcement learning and machine learning. A majority of the rich literature available for saddle-point optimization has focused on the offline setting. In this paper, we study nonstationary versions of stochastic, smooth, strongly-convex and strongly-concave saddle-point optimization problem, in both online (or first-order) and multi-point bandit (or zeroth-order) settings. We fir… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
7

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(11 citation statements)
references
References 20 publications
0
11
0
Order By: Relevance
“…We refer the reader to (Daskalakis et al, 2021) for a more thorough discussion on the literature. Several recent works start considering the problem of learning over a sequence of non-stationary payoffs under different structures, including zero-sum matrix games (Cardoso et al, 2019;Fiez et al, 2021), convex-concave games (Roy et al, 2019) and strongly monotone games (Duvocelle et al, 2021). For zero-sum games, (Fiez et al, 2021) focuses on the periodic case and proves divergence results for a class of learning algorithms; (Cardoso et al, 2019) is the closest to our work, but as mentioned, we argue that their proposed measure (NE-regret) is not always appropriate (see Section 3.1).…”
Section: Measurementioning
confidence: 99%
See 2 more Smart Citations
“…We refer the reader to (Daskalakis et al, 2021) for a more thorough discussion on the literature. Several recent works start considering the problem of learning over a sequence of non-stationary payoffs under different structures, including zero-sum matrix games (Cardoso et al, 2019;Fiez et al, 2021), convex-concave games (Roy et al, 2019) and strongly monotone games (Duvocelle et al, 2021). For zero-sum games, (Fiez et al, 2021) focuses on the periodic case and proves divergence results for a class of learning algorithms; (Cardoso et al, 2019) is the closest to our work, but as mentioned, we argue that their proposed measure (NE-regret) is not always appropriate (see Section 3.1).…”
Section: Measurementioning
confidence: 99%
“…Another reasonable one is the tracking error T t=1 ( x t −x * t 1 + y t −y * t 1 ) that directly measures the distance between (x t , y t ) and the equilibrium (x * t , y * t ) (assuming unique equilibrium for simplicity). This is considered in (Roy et al, 2019;Balasubramanian & Ghadimi, 2021) (for different problems). However, tracking error bounds are in fact not well studied even when A t is fixed -the best known results still depend on some problem-dependent constant that can be arbitrarily large (Daskalakis & Panageas, 2019;Wei et al, 2021).…”
Section: Duality Gapmentioning
confidence: 99%
See 1 more Smart Citation
“…These approaches typically opt for a tracking error metric. In the more general online saddle point problem, one seeks to find a sequence of strategy pairs that minimize a saddle point regret [28,46].…”
Section: Related Workmentioning
confidence: 99%
“…For convex-concave minimax optimization problems, there are some exisiting algorithms. For example, Roy et al [47] study zeroth-order Frank-Wolfe algorithms for strongly convex-strongly concave minimax optimization problems and provide non-asymptotic oracle complexity analysis. Beznosikov et al [4] present a zeroth-order saddle-point algorithm (zoSPA) with the total complexity of O ε −2 .…”
mentioning
confidence: 99%