2020
DOI: 10.48550/arxiv.2004.00603
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

No-Regret Learning Dynamics for Extensive-Form Correlated Equilibrium

Abstract: Recently, there has been growing interest around less-restrictive solution concepts than Nash equilibrium in extensive-form games, with significant effort towards the computation of extensive-form correlated equilibrium (EFCE) and extensive-form coarse correlated equilibrium (EFCCE). In this paper, we show how to leverage the popular counterfactual regret minimization (CFR) paradigm to induce simple no-regret dynamics that converge to the set of EFCEs and EFCCEs in an n-player general-sum extensive-form games.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
1

Year Published

2020
2020
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(7 citation statements)
references
References 7 publications
(9 reference statements)
0
6
1
Order By: Relevance
“…present the first line of results for learning Nash, CE, and CCE in general-sum Markov games; however their sample complexity scales with i≤m A i due to the model-based nature of their algorithm. Algorithms for learning CE in extensive-form games has been studied in (Celli et al, 2020), though we remark Markov games and extensive-form games are different frameworks and our results do not imply each other.…”
Section: Related Workcontrasting
confidence: 72%
“…present the first line of results for learning Nash, CE, and CCE in general-sum Markov games; however their sample complexity scales with i≤m A i due to the model-based nature of their algorithm. Algorithms for learning CE in extensive-form games has been studied in (Celli et al, 2020), though we remark Markov games and extensive-form games are different frameworks and our results do not imply each other.…”
Section: Related Workcontrasting
confidence: 72%
“…The theorem instead implies that the historical average of the CFR policies π t , which is the policy returned by the DREAM algorithm, converges to a Nash equilibrium at this rate in two-player zero-sum games. In n-player general-sum games, it also converges to an extensive-form coarse correlated equilibrium at this rate [8].…”
Section: D2 Bound On Dream's Regretmentioning
confidence: 97%
“…Later, Farina et al (2019) propose a min-max optimization formulation of EFCEs which can be solved by first-order methods. Celli et al (2020) and its extended version (Farina et al, 2021a) design the first uncoupled no-regret algorithm for computing EFCEs. Their algorithms are based on minimizing the trigger regret (first considered in Dudik and Gordon (2012); Gordon et al (2008)) via counterfactual regret decomposition (Zinkevich et al, 2007).…”
Section: Related Workmentioning
confidence: 99%
“…Equivalence between 1-EFCE and trigger definition of EFCE At the special case K = 1, our (exact) 1-EFCE is equivalent to the existing definition of EFCE based on trigger policies (Gordon et al, 2008;Celli et al, 2020), which defines an ε-approximate EFCE as any correlated policy π such that the following trigger gap is at most ε:…”
Section: Properties Of K-efcementioning
confidence: 99%
See 1 more Smart Citation