2020 IEEE International Conference on Robotics and Automation (ICRA) 2020
DOI: 10.1109/icra40945.2020.9197020
|View full text |Cite
|
Sign up to set email alerts
|

Scalable Multi-Task Imitation Learning with Autonomous Improvement

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 25 publications
(18 citation statements)
references
References 11 publications
0
18
0
Order By: Relevance
“…Recent works have focused on scaling up multitask imitation learning over unstructured data [8,26]. However, these approaches typically assume that tasks are specified to the agent at test time via mechanisms like one-hot task selectors [37,43], goal images [29,11,26], or target configurations of the state space [14]. While these types of conditioning are straightforward to provide in simulation, they are often impractical to provide in open-world settings.…”
Section: Introductionmentioning
confidence: 99%
“…Recent works have focused on scaling up multitask imitation learning over unstructured data [8,26]. However, these approaches typically assume that tasks are specified to the agent at test time via mechanisms like one-hot task selectors [37,43], goal images [29,11,26], or target configurations of the state space [14]. While these types of conditioning are straightforward to provide in simulation, they are often impractical to provide in open-world settings.…”
Section: Introductionmentioning
confidence: 99%
“…Multi-task Imitation Learning for Robotic Manipulation Our work falls under the broader category of imitation learning multiple robot manipulation tasks [47][24] [35]. The term "multi-task" has varying definitions across this space of literature.…”
Section: Related Work Imitation Learningmentioning
confidence: 99%
“…On the other hand, π S performed worse than the baseline in task T L . As result, π S may not be considered a zero-shot learner [Singh et al 2020, Oh et al 2017] in terms of generalizing for all 4 tasks defined, because it was not able to generalize to the unseen task T L . This answer RQ4.…”
Section: Policy Evaluation and Improvementmentioning
confidence: 99%