2021 IEEE International Conference on Robotics and Automation (ICRA) 2021
DOI: 10.1109/icra48506.2021.9560734
|View full text |Cite
|
Sign up to set email alerts
|

Dreaming: Model-based Reinforcement Learning by Latent Imagination without Reconstruction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
37
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 40 publications
(37 citation statements)
references
References 5 publications
0
37
0
Order By: Relevance
“…While previous work often found contrastive learning to be ineffective, we show that combining it with recurrent state space models makes it work. Recently, a contrastive variant of Dreamer (Okada & Taniguchi, 2020) has been proposed which shares the same motivation. Concurrent with our work, Nguyen et al ( 2021) explore a formulation similar to ours based on temporal predictive coding, but do not evaluate it on the difficult camera and color distractions we do here.…”
Section: Discussion and Related Workmentioning
confidence: 99%
“…While previous work often found contrastive learning to be ineffective, we show that combining it with recurrent state space models makes it work. Recently, a contrastive variant of Dreamer (Okada & Taniguchi, 2020) has been proposed which shares the same motivation. Concurrent with our work, Nguyen et al ( 2021) explore a formulation similar to ours based on temporal predictive coding, but do not evaluate it on the difficult camera and color distractions we do here.…”
Section: Discussion and Related Workmentioning
confidence: 99%
“…Relying on image reconstruction can however lead to vulnerability to visual noise: to overcome this limitation Okada and Taniguchi [33] and Zhang et al [43] forgo the decoder network, while the latter proposes to rely on the notion of bisimilarity to learn meaningful representations. Similarly, Gelada et al [16] only learn to predict rewards and action-conditional state distributions, but only study this task as an additional loss to model-free reinforcement learning methods.…”
Section: Related Workmentioning
confidence: 99%
“…World models [1] are a potential approach to achieving the visual servoing of robots in the industry. The world models, which are equipped with compact latent representation models and latent forward dynamics, efficiently predict future trajectories and rewards, allowing us to acquire model predictive controllers [2]- [4] and policies learned by model-based reinforcement learning [5]- [7]. In addition, world modes have various valuable properties for industrial applications, such as transferability to new tasks [8], unsupervised exploration [9], generalization from offline datasets [10], and explainability [11].…”
Section: Introductionmentioning
confidence: 99%
“…DreamerV2 [6] is a leading type of world model based reinforcement learning that achieved human-level performance on the Atari benchmark. Unlike previous world models [2], [3], [7], including Dreamer [5] (the earlier version of Dream-erV2), this method uses discrete world models in which discrete random variables represent latent states. A motivation to introduce discrete representation is that categorical distributions can naturally capture multimodal uncertainty of stochastic state transitions.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation