2022
DOI: 10.48550/arxiv.2203.11409
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Primer on Maximum Causal Entropy Inverse Reinforcement Learning

Abstract: Inverse Reinforcement Learning (IRL) algorithms [17,1] infer a reward function that explains demonstrations provided by an expert acting in the environment. Maximum Causal Entropy (MCE) IRL [31,29] is currently the most popular formulation of IRL, with numerous extensions [5,8,21]. In this tutorial, we present a compressed derivation of MCE IRL and the key results from contemporary implementations of MCE IRL algorithms. We hope this will serve both as an introductory resource for those new to the field, and as… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 8 publications
0
1
0
Order By: Relevance
“…Although the original feature matching objective only considers reward functions that are linear in φ, state-of-the-art methods are scalable to complex problems by parameterizing R with a deep neural network (Wulfmeier et al, 2015;Ho & Ermon, 2016). For an elaborate introduction to MCE-IRL we refer the reader to the work of Gleave & Toyer (2022).…”
Section: Inverse Reinforcement Learningmentioning
confidence: 99%
“…Although the original feature matching objective only considers reward functions that are linear in φ, state-of-the-art methods are scalable to complex problems by parameterizing R with a deep neural network (Wulfmeier et al, 2015;Ho & Ermon, 2016). For an elaborate introduction to MCE-IRL we refer the reader to the work of Gleave & Toyer (2022).…”
Section: Inverse Reinforcement Learningmentioning
confidence: 99%