2020
DOI: 10.1007/978-3-030-55754-6_7
|View full text |Cite
|
Sign up to set email alerts
|

Strengthening Deterministic Policies for POMDPs

Abstract: The synthesis problem for partially observable Markov decision processes (POMDPs) is to compute a policy that satisfies a given specification. Such policies have to take the full execution history of a POMDP into account, rendering the problem undecidable in general. A common approach is to use a limited amount of memory and randomize over potential choices. Yet, this problem is still NP-hard and often computationally intractable in practice. A restricted problem is to use neither history nor randomization, yi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
2
2
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 35 publications
(40 reference statements)
0
2
0
Order By: Relevance
“…Quantitative variants of reach-avoid specifications have gained attention in, e.g., [11,28,40]. Other approaches restrict themselves to simple policies [3,33,45,58]. Wang et al [55] use an iterative Satisfiability Modulo Theories (SMT) [6] approach for quantitative finite-horizon specifications, which requires computing beliefs.…”
Section: Contributions Our Paper Makes Four Contributions: (1)mentioning
confidence: 99%
“…Quantitative variants of reach-avoid specifications have gained attention in, e.g., [11,28,40]. Other approaches restrict themselves to simple policies [3,33,45,58]. Wang et al [55] use an iterative Satisfiability Modulo Theories (SMT) [6] approach for quantitative finite-horizon specifications, which requires computing beliefs.…”
Section: Contributions Our Paper Makes Four Contributions: (1)mentioning
confidence: 99%
“…The use of [63] of game-based abstraction leads to non-randomized controllers. Winterer et al [64] support a finite set of uniform randomizations. In contrast, we consider an infinite combination of possibilities.…”
Section: Related Workmentioning
confidence: 99%