2010
DOI: 10.1609/aaai.v24i1.7571
|View full text |Cite
|
Sign up to set email alerts
|

Learning Methods to Generate Good Plans: Integrating HTN Learning and Reinforcement Learning

Abstract: We consider how to learn Hierarchical Task Networks (HTNs) for planning problems in which both the quality of solution plans generated by the HTNs and the speed at which those plans are found is important. We describe an integration of HTN Learning with Reinforcement Learning to both learn methods by analyzing semantic annotations on tasks and to produce estimates of the expected values of the learned methods by performing Monte Carlo updates. We performed an experiment in which plan quality was inversely re… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
10
0

Year Published

2011
2011
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 18 publications
(10 citation statements)
references
References 8 publications
(6 reference statements)
0
10
0
Order By: Relevance
“…While the subgoal structure of a domain is important for handcrafting effective hierarchical task networks, HTNs do not actually encode subgoal structures but general top-down solving strategies where (non-primitive) methods decompose into other methods (Erol et al, 1994;Nau et al, 1999;Georgievski & Aiello, 2015). Techniques for learning HTNs usually appeal to annotated traces that convey the intended decompositions (Hogg et al, 2008;Zhuo et al, 2009). Other methods for deriving hierarchical decompositions in planning include precondition relaxations (Sacerdoti, 1974) and causal graphs (Knoblock, 1994).…”
Section: Related Workmentioning
confidence: 99%
“…While the subgoal structure of a domain is important for handcrafting effective hierarchical task networks, HTNs do not actually encode subgoal structures but general top-down solving strategies where (non-primitive) methods decompose into other methods (Erol et al, 1994;Nau et al, 1999;Georgievski & Aiello, 2015). Techniques for learning HTNs usually appeal to annotated traces that convey the intended decompositions (Hogg et al, 2008;Zhuo et al, 2009). Other methods for deriving hierarchical decompositions in planning include precondition relaxations (Sacerdoti, 1974) and causal graphs (Knoblock, 1994).…”
Section: Related Workmentioning
confidence: 99%
“…While the subgoal structure of a domain is important for handcrafting effective hierarchical task networks, HTNs do not actually encode subgoal structures but general top-down strategies where (non-primitive) methods decompose into other methods (Erol, Hendler, & Nau, 1994;Nau, Cao, Lotem, & Munoz-Avila, 1999;Georgievski & Aiello, 2015). Techniques for learning HTNs usually appeal to annotated traces that convey the intended decompositions (Hogg, Munoz-Avila, & Kuter, 2008;Zhuo, Hu, Hogg, Yang, & Munoz-Avila, 2009), and other methods for deriving hierarchical decompositions in planning include precondition relaxations (Sacerdoti, 1974) and causal graphs (Knoblock, 1994).…”
Section: Related Workmentioning
confidence: 99%
“…However, it still requires demonstrations to be provided as decomposition trees, divided by which higher-level goal is being pursued at any given point. (Hogg, Kuter, and Munoz-Avila 2010) integrates reinforcement learning into the HTN-MAKER framework, in order to ascertain values for the learned methods to decide which are more likely to be useful in any given situation. Using this setup, the authors manage to improve the rate at which the HTN model learns from demonstration, requiring less examples before achieving competency.…”
Section: Related Workmentioning
confidence: 99%