2018
DOI: 10.1561/2300000053
|View full text |Cite
|
Sign up to set email alerts
|

An Algorithmic Perspective on Imitation Learning

Abstract: As robots and other intelligent agents move from simple environments and problems to more complex, unstructured settings, manually programming their behavior has become increasingly challenging and expensive. Often, it is easier for a teacher to demonstrate a desired behavior rather than attempt to manually engineer it. This process of learning from demonstrations, and the study of algorithms to do so, is called imitation learning. This work provides an introduction to imitation learning. It covers the underly… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
302
0
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 385 publications
(329 citation statements)
references
References 138 publications
2
302
0
2
Order By: Relevance
“…These methods attempt to generalize from previously observed interactions to predict multi-agent behavior in new situations. Forecasting is related to Imitation Learning [25], which learns a model to mimic demonstrated behavior. In contrast to some Imitation Learning methods, e.g.…”
Section: Related Workmentioning
confidence: 99%
“…These methods attempt to generalize from previously observed interactions to predict multi-agent behavior in new situations. Forecasting is related to Imitation Learning [25], which learns a model to mimic demonstrated behavior. In contrast to some Imitation Learning methods, e.g.…”
Section: Related Workmentioning
confidence: 99%
“…Since we used 3 features, θ's dimensionality was 3, leading to a possible set Θ equivalent to the 3-fold Cartesian product of the values above. After normalizing to norm 1, we were left with 19 unique θ vectors in Θ, weighing the three features in different proportions, as shown in Figures 3,7,8,9,and 10. Our discretization scheme ensured an approximately uniform sampling on the positive quadrant of the unit sphere.…”
Section: Appendix a Practical Considerationsmentioning
confidence: 99%
“…Most similar to our setting is inverse reinforcement learning (IRL), an instance of LfD where the robot learns the correct reward function from human demonstrations [3], [4]. Prior works on IRL generally assume that every human has a single, fixed teaching strategy [5]: the human teaches by providing optimal demonstrations, and any sub-optimal human behavior is interpreted as noise [6]- [9]. Alternatively, robots can also learn about the human while learning from that human.…”
Section: Related Work a Robots Learning From Humansmentioning
confidence: 99%
“…To compare our learning with strategy uncertainty against the state-of-the-art in a realistic problem setting, we performed a simulated user study. We here consider an instance of inverse reinforcement learning (IRL): the human demonstrates a policy, and the robot attempts to infer the human's reward function from that demonstrated policy [3]- [5]. Unlike the example in Section IV-C, now θ * (the human's reward parameters) and φ * (the human's demonstration strategy) lie in continuous spaces.…”
Section: Robot Learning Simulationsmentioning
confidence: 99%