2009 IEEE/RSJ International Conference on Intelligent Robots and Systems 2009
DOI: 10.1109/iros.2009.5354013
|View full text |Cite
|
Sign up to set email alerts
|

Bayesian reinforcement learning in continuous POMDPs with gaussian processes

Abstract: Partially Observable Markov Decision Processes (POMDPs) provide a rich mathematical model to handle realworld sequential decision processes but require a known model to be solved by most approaches. However, mainstream POMDP research focuses on the discrete case and this complicates its application to most realistic problems that are naturally modeled using continuous state spaces. In this paper, we consider the problem of optimal control in continuous and partially observable environments when the parameters … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
25
0

Year Published

2013
2013
2018
2018

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 26 publications
(26 citation statements)
references
References 13 publications
0
25
0
Order By: Relevance
“…Planning under uncertainty and mapping are combined in approaches that attempt to simultaneously capture the environment model and optimize the policy [33,139]. Up to now, those approaches are only valid for relatively small problems.…”
Section: Bibliographical Notesmentioning
confidence: 99%
“…Planning under uncertainty and mapping are combined in approaches that attempt to simultaneously capture the environment model and optimize the policy [33,139]. Up to now, those approaches are only valid for relatively small problems.…”
Section: Bibliographical Notesmentioning
confidence: 99%
“…Another approach proposed by Dallaire et al [2009] allows flexibility over the choice of transition function. Here the transition and reward functions are defined by:…”
Section: Extensions To Continuous Mdpsmentioning
confidence: 99%
“…In contrast to previous work on continuous POMDPs (e.g., [41]), we are focused on large structured action spaces (i.e., all possible routings of the cooperative vehicles).…”
Section: Related Workmentioning
confidence: 99%