2020
DOI: 10.48550/arxiv.2011.14381
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Self-supervised Visual Reinforcement Learning with Object-centric Representations

Abstract: Autonomous agents need large repertoires of skills to act reasonably on new tasks that they have not seen before. However, acquiring these skills using only a stream of high-dimensional, unstructured, and unlabeled observations is a tricky challenge for any autonomous agent. Previous methods have used variational autoencoders to encode a scene into a low-dimensional vector that can be used as a goal for an agent to discover new skills. Nevertheless, in compositional/multiobject environments it is difficult to … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 11 publications
0
2
0
Order By: Relevance
“…If such benefits can be extended across agents, fair comparisons in RL should require listing the amount of compute resources used for novel methods. This is especially important for compute-intensive unsupervised methods [1,38,43,58,70,72] or model-based learning [4,12,20,27,41]. There are two common approaches for ensuring fair comparisons: using a standard architecture across algorithms or listing the amount of compute/memory consumption and compare methods on this basis.…”
Section: Discussionmentioning
confidence: 99%
“…If such benefits can be extended across agents, fair comparisons in RL should require listing the amount of compute resources used for novel methods. This is especially important for compute-intensive unsupervised methods [1,38,43,58,70,72] or model-based learning [4,12,20,27,41]. There are two common approaches for ensuring fair comparisons: using a standard architecture across algorithms or listing the amount of compute/memory consumption and compare methods on this basis.…”
Section: Discussionmentioning
confidence: 99%
“…However, collecting real interaction trajectories is very timeconsuming, while physics parameters in real situations can only be approximated, making the model-based methods hard to apply. A model-free method like RL can be used to get actions directly from the ground truth state (Peng et al, 2018), or raw pixel images (Zadaianchuk et al, 2020). However, generalization of the model to different manipulation objects is hard for ground truth input, while extracting useful features such as object shape, size, and the robot's relative position efficiently from raw images for a subsequent policy network is always tricky.…”
Section: Planar Object Pushingmentioning
confidence: 99%