2020
DOI: 10.48550/arxiv.2010.04296
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

CausalWorld: A Robotic Manipulation Benchmark for Causal Structure and Transfer Learning

Abstract: Despite recent successes of reinforcement learning (RL), it remains a challenge for agents to transfer learned skills to related environments. To facilitate research addressing this problem, we propose CausalWorld, a benchmark for causal structure and transfer learning in a robotic manipulation environment. The environment is a simulation of an open-source robotic platform, hence offering the possibility of sim-to-real transfer. Tasks consist of constructing 3D shapes from a given set of blocks -inspired by ho… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
34
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 22 publications
(34 citation statements)
references
References 31 publications
(39 reference statements)
0
34
0
Order By: Relevance
“…Benchmark datasets play an important role in developing machine learning methodologies. Examples include ImageNet (Deng et al, 2009) or MSCOCO (Lin et al, 2014) for computer vision, as well as cart-pole (Barto et al, 1983) or reinforcement learning (Ahmed et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Benchmark datasets play an important role in developing machine learning methodologies. Examples include ImageNet (Deng et al, 2009) or MSCOCO (Lin et al, 2014) for computer vision, as well as cart-pole (Barto et al, 1983) or reinforcement learning (Ahmed et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…We evaluate and analyze our proposed MOC DRL on the CausalWorld [Ahmed et al, 2020], as this environment enables us to easily design and test different types of curricula in a fine-grained manner. It should be noted that we do not utilize any causal elements of the environment.…”
Section: Methodsmentioning
confidence: 99%
“…One of the biggest challenges in reinforcement learning (RL) is the brittleness of trained agents to distribution shifts in the environment. Recent studies have developed benchmarks to quantify the generalization performance of RL agents in out-of-distribution environments [10,11,12]. Indeed, this problem is particularly relevant to the field of robot learning where policies are often trained in simulation and directly transferred to hardware, resulting in an OOD deployment of the policy due to the mismatch between the simulator and the real world.…”
Section: Related Workmentioning
confidence: 99%