2022
DOI: 10.1613/jair.1.12889
|View full text |Cite
|
Sign up to set email alerts
|

A Survey of Opponent Modeling in Adversarial Domains

Abstract: Opponent modeling is the ability to use prior knowledge and observations in order to predict the behavior of an opponent. This survey presents a comprehensive overview of existing opponent modeling techniques for adversarial domains, many of which must address stochastic, continuous, or concurrent actions, and sparse, partially observable payoff structures. We discuss all the components of opponent modeling systems, including feature extraction, learning algorithms, and strategy abstractions. These discussions… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(8 citation statements)
references
References 113 publications
0
5
0
Order By: Relevance
“…Our work is related to these in the sense that we assume opponent models to be given (called "type-based reasoning" by Albrecht and Stone (2018, Section 4.2)). However, an important stream of work also studies the learning of opponent models; we refer the reader to the survey by Nashed and Zilberstein (2022).…”
Section: Related Workmentioning
confidence: 99%
“…Our work is related to these in the sense that we assume opponent models to be given (called "type-based reasoning" by Albrecht and Stone (2018, Section 4.2)). However, an important stream of work also studies the learning of opponent models; we refer the reader to the survey by Nashed and Zilberstein (2022).…”
Section: Related Workmentioning
confidence: 99%
“…Opponent modeling is the problem of estimating the properties of an opponent (Nashed & Zilberstein, 2022). Much previous work on this topic has been done in imperfect information games like poker (Billings et al, 1998;Bard et al, 2015Bard et al, , 2013Davis et al, 2014), but this work focuses on strategic characteristics and limitations of the opponents, and the domains do not include execution uncertainty.…”
Section: Opponent Modelingmentioning
confidence: 99%
“…Policy reconstruction methods predict agent policies from environment observations [2], which has been shown to be beneficial in collaborative [14], competitive [3] and mixed settings [15]. Deep Reinforcement Opponent Modelling (DRON) was one of the first works combining deep RL (DRL) with opponent modelling [16].…”
Section: A Opponent Modelling In Drlmentioning
confidence: 99%
“…, by freezing a copy of a PPO [9] agent trained under δ = 0-Uniform self-play [6] after 200k, 400k and 600k episodes respectively. [0,1], [1,2], [2,3], [3,4] Table I shows the hyperparameters used to train the test agents. No formal hyperparameter sweep was conducted, and the final values were chosen after a few manual trials.…”
Section: A Training and Benchmarking Opponentsmentioning
confidence: 99%
See 1 more Smart Citation