2013
DOI: 10.1109/tciaig.2013.2248086
|View full text |Cite
|
Sign up to set email alerts
|

Incentive Learning in Monte Carlo Tree Search

Abstract: Monte Carlo tree search (MCTS) is a search paradigm that has been remarkably successful in computer games like Go. It uses Monte Carlo simulation to evaluate the values of nodes in a search tree. The node values are then used to select the actions during subsequent simulations. The performance of MCTS heavily depends on the quality of its default policy, which guides the simulations beyond the search tree. In this paper, we propose an MCTS improvement, called incentive learning, which learns the default policy… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 13 publications
(18 reference statements)
0
2
0
Order By: Relevance
“…The classical RAVE approach (see Sect. 2.2) has been extended by Kao et al (2013). They introduce the RIDE method (Rapid Incentive Difference Evaluation) where the default MCTS policy is updated by using differences (9) between action values for the same state s.…”
Section: Policy Updatementioning
confidence: 99%
“…The classical RAVE approach (see Sect. 2.2) has been extended by Kao et al (2013). They introduce the RIDE method (Rapid Incentive Difference Evaluation) where the default MCTS policy is updated by using differences (9) between action values for the same state s.…”
Section: Policy Updatementioning
confidence: 99%
“…The classical RAVE approach (see Section 2.2) has been extended by Kao et al (2013). They introduce the RIDE method (Rapid Incentive Difference Evaluation) where the default MCTS policy is updated by using differences (9) between action values for the same state s.…”
Section: Policy Updatementioning
confidence: 99%