2020
DOI: 10.48550/arxiv.2001.08174
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Robust Policy Synthesis for Uncertain POMDPs via Convex Optimization

Abstract: We study the problem of policy synthesis for uncertain partially observable Markov decision processes (uPOMDPs). The transition probability function of uPOMDPs is only known to belong to a so-called uncertainty set, for instance in the form of probability intervals. Such a model arises when, for example, an agent operates under information limitation due to imperfect knowledge about the accuracy of its sensors. The goal is to compute a policy for the agent that is robust against all possible probability distri… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 3 publications
(5 reference statements)
0
1
0
Order By: Relevance
“…One of the general frameworks of the path planning is to describe it as a partially observable Markov decision process (POMDP) [21][22][23]. In this framework, plenty of methods are searched [24][25][26][27].…”
Section: Introduction 11 Backgroundmentioning
confidence: 99%
“…One of the general frameworks of the path planning is to describe it as a partially observable Markov decision process (POMDP) [21][22][23]. In this framework, plenty of methods are searched [24][25][26][27].…”
Section: Introduction 11 Backgroundmentioning
confidence: 99%