Robotics: Science and Systems XVI 2020
DOI: 10.15607/rss.2020.xvi.061
|View full text |Cite
|
Sign up to set email alerts
|

GTI: Learning to Generalize across Long-Horizon Tasks from Human Demonstrations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
44
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 55 publications
(44 citation statements)
references
References 13 publications
0
44
0
Order By: Relevance
“…During the demonstration phase, it parses the recorded video to extract the 6 DoF motion trajectory of the manipulated object in the receptacle's coordinate frame. Compared to learning directly in the image pixel space [33], this approach disentangles the object of interest from the background and represents the extracted trajectory independent of any specific scene configuration. This enables the representation to generalize to novel environments, where the initial object and receptacle placement might differ from the demonstration.…”
Section: B Model-free 6 Dof Object Motion Trackingmentioning
confidence: 99%
See 1 more Smart Citation
“…During the demonstration phase, it parses the recorded video to extract the 6 DoF motion trajectory of the manipulated object in the receptacle's coordinate frame. Compared to learning directly in the image pixel space [33], this approach disentangles the object of interest from the background and represents the extracted trajectory independent of any specific scene configuration. This enables the representation to generalize to novel environments, where the initial object and receptacle placement might differ from the demonstration.…”
Section: B Model-free 6 Dof Object Motion Trackingmentioning
confidence: 99%
“…The demonstration video and an example successful robot execution are shown in the first page's figure (top). This task represents commonly considered pickand-place tasks in prior efforts on visual imitation learning [22,34,23,46,60,33,50]. This task is the simplest among the ones considered here but still requires accurate orientation reasoning for stable placement due to the batteries' shape, which is usually long and thin.…”
Section: Battery Standing Taskmentioning
confidence: 99%
“…simulation vs. reality). Current imitation learning approaches from pixels are mainly tested within a single domain [23,24,17,38], domains with a limited gap [2,15,34], or require manually labelled data [25]. For example, the third-person imitation work by Stadie et al [34] deals with known actions, observations from the same simulated embodiment, but with a different viewpoint.…”
Section: Related Workmentioning
confidence: 99%
“…In the imitation learning (IL) context, the control policy is trained using human demonstrations [23]- [27]. This approach provides a safe and efficient way to learn complex skills; however, it is difficult to generalize towards the state space outside the collected demonstration.…”
Section: Related Workmentioning
confidence: 99%
“…Generalization: So far, previous studies tackled the problem of generalization. Several frameworks were proposed to achieve generalizable IL policy in terms of executable tasks and domains [26], [27]. Ebert et al [31] collected a dataset to boost cross-task and cross-domain generalization.…”
Section: Related Workmentioning
confidence: 99%