2019
DOI: 10.48550/arxiv.1910.13038
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning to Predict Without Looking Ahead: World Models Without Forward Prediction

Abstract: Much of model-based reinforcement learning involves learning a model of an agent's world, and training an agent to leverage this model to perform a task more efficiently. While these models are demonstrably useful for agents, every naturally occurring model of the world of which we are aware-e.g., a brain-arose as the byproduct of competing evolutionary pressures for survival, not minimization of a supervised forward-predictive loss via gradient descent. That useful models can arise out of the messy and slow o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 43 publications
(46 reference statements)
0
1
0
Order By: Relevance
“…One of the most challenging aspects of MBRL is that the model that is acquired must be sophisticated enough to capture the dynamics of the environment while being flexible and robust to potential errors. Some research has focused on producing faithful reconstructions (Kaiser et al 2020; while others have opted instead on learning to predict specific aspects of the RL framework (Schrittwieser et al 2020) or the environment (Freeman, Metz, and Ha 2019) that are thought to be necessary to perform a particular task.…”
Section: State Space Models In Model-based Reinforcement Learningmentioning
confidence: 99%
“…One of the most challenging aspects of MBRL is that the model that is acquired must be sophisticated enough to capture the dynamics of the environment while being flexible and robust to potential errors. Some research has focused on producing faithful reconstructions (Kaiser et al 2020; while others have opted instead on learning to predict specific aspects of the RL framework (Schrittwieser et al 2020) or the environment (Freeman, Metz, and Ha 2019) that are thought to be necessary to perform a particular task.…”
Section: State Space Models In Model-based Reinforcement Learningmentioning
confidence: 99%