2018
DOI: 10.48550/arxiv.1808.00928
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning Actionable Representations from Visual Observations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 0 publications
0
3
0
Order By: Relevance
“…Self-supervised learning from sequences. Previous work in contrastive learning for sequential data often leverages a slowness assumption to use nearby samples as positive examples and farther samples as negative examples (Oord et al, 2018;Sermanet et al, 2018;Dwibedi et al, 2019;Le-Khac et al, 2020;Banville et al, 2020). Contrastive predictive coding (CPC) (Oord et al, 2018) builds upon the idea of temporal contrastive learning by building an AR-model that predicts future points given previous observed timesteps.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Self-supervised learning from sequences. Previous work in contrastive learning for sequential data often leverages a slowness assumption to use nearby samples as positive examples and farther samples as negative examples (Oord et al, 2018;Sermanet et al, 2018;Dwibedi et al, 2019;Le-Khac et al, 2020;Banville et al, 2020). Contrastive predictive coding (CPC) (Oord et al, 2018) builds upon the idea of temporal contrastive learning by building an AR-model that predicts future points given previous observed timesteps.…”
Section: Related Workmentioning
confidence: 99%
“…Temporal shift: As in previous work in temporal contrastive learning (Oord et al, 2018;Sermanet et al, 2018;Dwibedi et al, 2019;Le-Khac et al, 2020;Banville et al, 2020), we can use nearby samples as positive examples for one another.…”
Section: B3 Augmentations For Neural Datamentioning
confidence: 99%
“…A third way we exploit viewpoint changes is for multiple-view self-supervised representation learning. The ability to observe different views of an object or a scene has been used in prior work ( [21], [23], [26], [27]) to learn low-dimensional state representations without human annotation. Efficient encoding of object and scene properties from high-dimensional images is essential for vision-based manipulation; we utilize Generative Query Networks [27] for this purpose.…”
Section: Introductionmentioning
confidence: 99%