2022 IEEE 61st Conference on Decision and Control (CDC) 2022
DOI: 10.1109/cdc51059.2022.9992950
|View full text |Cite
|
Sign up to set email alerts
|

Improved Rates for Derivative Free Gradient Play in Strongly Monotone Games

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

1
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 13 publications
1
3
0
Order By: Relevance
“…By employing a barrier-based method, Lin et al [15] improved the convergence rate for strongly monotone games from đť‘‚ (1/𝑡 1/3 ) to đť‘‚ (1/𝑡 1/2 ). Similar convergence rates have also been reported in [16], [17], [18]. Huang et al [19] developed two bandit learning algorithms by integrating residual pseudo-gradient estimates into singlecall extra-gradient schemes that ensure a.s. convergence to critical points of pseudo-monotone plus games.…”
Section: Introductionsupporting
confidence: 60%
See 1 more Smart Citation
“…By employing a barrier-based method, Lin et al [15] improved the convergence rate for strongly monotone games from đť‘‚ (1/𝑡 1/3 ) to đť‘‚ (1/𝑡 1/2 ). Similar convergence rates have also been reported in [16], [17], [18]. Huang et al [19] developed two bandit learning algorithms by integrating residual pseudo-gradient estimates into singlecall extra-gradient schemes that ensure a.s. convergence to critical points of pseudo-monotone plus games.…”
Section: Introductionsupporting
confidence: 60%
“…Contributions: In this work, we develop a bandit online learning algorithm and establish the a.s. convergence of the generated sequence of play under the regularity condition that the game is merely coherent, which is broader and more general than the games investigated in [14], [15], [16], [17], [18]. The proposed algorithm leverages the optimistic mirror descent (OMD) [30], [31] and a single-call extragradient scheme as the backbone, which allows us to deal with the absence of strict coherence and reduces the query cost induced by the extra step.…”
Section: Introductionmentioning
confidence: 99%
“…The analysis of Mertikopoulos and Zhou [40] is subsequently extended by Bravo et al [13] to learning with payoff-based, "bandit feedback"-that is, when players observe only the payoff of the action that they played. At around the same time, Tatarenko and Kamgarpour [63,64] use a Tikhonov regularization approach to obtain a series of comparable results for "merely monotone" games (i.e., monotone games that are not necessarily strictly monotone), whereas more recently, Drusvyatskiy and Ratliff [21] improve the rate of convergence in strongly monotone games to O(1=T 1=2 ). Finally, in a very recent paper, Bervoets et al [9] use stochastic approximation methodologies to prove the convergence of a payoff-based, dampened gradient approximation scheme in two other classes of one-dimensional concave games: games with strategic complements and ordinal potential games with isolated equilibria.…”
Section: Related Workmentioning
confidence: 99%
“…Assuming bandit feedback, i.e., agents only have access to zeroth-order oracles, [8] shows Nash equilibrium convergence for strongly monotone games, which is a special class of convex games. The convergence rate of the zeroth-order method in [8] is further improved in [30] relying on the additional assumption that the Jacobian of the gradient function is Lipschitz continuous. Common in these works is that the agents perform symmetric updates using the same kind of information feedback.…”
Section: Introductionmentioning
confidence: 99%