2022
DOI: 10.1609/aaai.v36i5.20457
|View full text |Cite
|
Sign up to set email alerts
|

Coordinating Followers to Reach Better Equilibria: End-to-End Gradient Descent for Stackelberg Games

Abstract: A growing body of work in game theory extends the traditional Stackelberg game to settings with one leader and multiple followers who play a Nash equilibrium. Standard approaches for computing equilibria in these games reformulate the followers' best response as constraints in the leader's optimization problem. These reformulation approaches can sometimes be effective, but make limiting assumptions on the followers' objectives and the equilibrium reached by followers, e.g., uniqueness, optimism, or pessimism. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(10 citation statements)
references
References 39 publications
0
10
0
Order By: Relevance
“…The incentive design could be reformulated as the Stackelberg game by treating the incentive design objective as the leading player. [86] proposed a gradient-decent algorithm to find NE for the Stackelberg bandit problem. [95] propose a value iteration method for solving the Stackelberg Markov game with convergence guarantee.…”
Section: Related Workmentioning
confidence: 99%
“…The incentive design could be reformulated as the Stackelberg game by treating the incentive design objective as the leading player. [86] proposed a gradient-decent algorithm to find NE for the Stackelberg bandit problem. [95] propose a value iteration method for solving the Stackelberg Markov game with convergence guarantee.…”
Section: Related Workmentioning
confidence: 99%
“…Theorem 3: Let the Stackelberg game be defined as in (4) and its auxiliary best-response optimization problem as (19). Let the Assumptions 1, 2, 3 and 4 hold and {π t } be the sequence generated by the projected gradient descent method defined by equations ( 11), (12) and (13). Then it holds that lim t→+∞ J L (•, π t+1 ) − J L (•, π t ) = 0 , and every limit point of {π t } is stationary.…”
Section: Convergence Of the Algorithmmentioning
confidence: 99%
“…The proposed semi-decentralized approach leverages the existence of a Central aggregator typically required for decentralized computation of the Nash equilibrium of the game played between the followers [14]- [16]. Inspired by [13], [17], we show how the followers can locally compute the Jacobians by the means of Implicit function theorem, in an attempt to estimate the influence of the leader's decision variable on the attained Nash equilibrium of the followers. By designing an iterative procedure that concerns the central aggregator and the leading player and guarantees local improvement of the leader's objective at each iteration, we can provide formal convergence guarantees.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations