Proceedings of the 13th International Conference on the Foundations of Digital Games 2018
DOI: 10.1145/3235765.3235812
|View full text |Cite
|
Sign up to set email alerts
|

A hybrid search agent in pommerman

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
14
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 19 publications
(14 citation statements)
references
References 5 publications
0
14
0
Order By: Relevance
“…Tree-based techniques such as Monte Carlo Tree Search (MCTS) have been shown to perform well in Pommerman Resnick et al (2020); Osogami and Takahashi (2019); Zhou et al (2018). However, they require much more computational infrastructure than pure RL, in addition to significant human effort for evaluating the trajectories.…”
Section: Related Workmentioning
confidence: 99%
“…Tree-based techniques such as Monte Carlo Tree Search (MCTS) have been shown to perform well in Pommerman Resnick et al (2020); Osogami and Takahashi (2019); Zhou et al (2018). However, they require much more computational infrastructure than pure RL, in addition to significant human effort for evaluating the trajectories.…”
Section: Related Workmentioning
confidence: 99%
“…The unique challenges in Pommerman have attracted many researchers to this environment. Their approaches can be broadly categorized into model-free RL [16,17,13,14,18] and tree-searchbased-RL [19,20,11,21,12]. In addition, [22] is an excellent review of Pommerman, its practical implications, and its limitations.…”
Section: Related Workmentioning
confidence: 99%
“…In addition, [22] is an excellent review of Pommerman, its practical implications, and its limitations. A comparison of search techniques including MCTS, breadth-first, and flat Monte Carlo [20] shows that in the fully observable FFA mode, MCTS is able to beat simpler and hand-crafted solutions. An extension of this study [19] called Rolling Horizon Evolutionary Algorithm (RHEA) concludes that the more offensive strategies (like RHEA with a high rate of bomb placing) are normally also riskier, due to inadvertent suicides 3 .…”
Section: Related Workmentioning
confidence: 99%
“…Their experiment in Pommerman shows that Backplay provides significant gains in sample complexity with a stark advantage in sparse reward setting. Hybrid Search Agent [25] focused on search-based methods in Pommerman with resource-intensive forward models. Their result shows that heuristic agent using depth-limited tree search can slightly outperform hand-made heuristics.…”
Section: Action Spacementioning
confidence: 99%