2012
DOI: 10.1007/978-3-642-31368-4_26
|View full text |Cite
|
Sign up to set email alerts
|

Adapting Strategies to Opponent Models in Incomplete Information Games: A Reinforcement Learning Approach for Poker

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
3
0
1

Year Published

2013
2013
2022
2022

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 13 publications
(4 citation statements)
references
References 6 publications
0
3
0
1
Order By: Relevance
“…In addition, opponent modeling can be adopted in multi-agent reinforcement learning problems where RL agents are designed to consider the learning of other agents in the environment when updating their own policies [36]. Another promising solution is to mimic human players by combining opponent models used by expert players and reinforcement learning [37]. All the above works demonstrate that combining opponent modeling with reinforcement learning is beneficial to achieve performance gain in multi-agent imperfect-information games, which also inspires this work.…”
Section: B Opponent Modeling For Gamesmentioning
confidence: 99%
“…In addition, opponent modeling can be adopted in multi-agent reinforcement learning problems where RL agents are designed to consider the learning of other agents in the environment when updating their own policies [36]. Another promising solution is to mimic human players by combining opponent models used by expert players and reinforcement learning [37]. All the above works demonstrate that combining opponent modeling with reinforcement learning is beneficial to achieve performance gain in multi-agent imperfect-information games, which also inspires this work.…”
Section: B Opponent Modeling For Gamesmentioning
confidence: 99%
“…Outro exemplo AKI-REALBOT onde foi utilizado o algoritmo de Monte-Carlo [5] de forma a explorar a estratégia do adversário. Outras técnicas utilizadas foram por exemplo a criação de agentes baseados em aprendizagem por reforço [6].…”
Section: Trabalho Relacionadounclassified
“…In [12], instead of the average, the degree of similarity was measured through the Euclidean distance between the game features. The Monte Carlo Search Tree algorithm [14] and reinforcement learning approaches [15] are other techniques that were successfully applied to the domain of Computer Poker. One should not also forget some work done in opponent modeling techniques, namely [16].…”
Section: One Great Breakthrough In the Domain Of Computer Poker And Other Extensive-form Games Research Was The Development Of The Countementioning
confidence: 99%