2020
DOI: 10.1007/978-3-030-45190-5_19
|View full text |Cite
|
Sign up to set email alerts
|

Simple Strategies in Multi-Objective MDPs

Abstract: We consider the verification of multiple expected reward objectives at once on Markov decision processes (MDPs). This enables a trade-off analysis among multiple objectives by obtaining a Pareto front. We focus on strategies that are easy to employ and implement. That is, strategies that are pure (no randomization) and have bounded memory. We show that checking whether a point is achievable by a pure stationary strategy is NP-complete, even for two objectives, and we provide an MILP encoding to solve the corre… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
21
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1
1

Relationship

4
3

Authors

Journals

citations
Cited by 22 publications
(22 citation statements)
references
References 38 publications
0
21
0
Order By: Relevance
“…• A comprehensive overview of strategy (Section 3, Table 1) and computational complexity (Section 4, Table 2) of disjunctive reachability-safety queries in stochastic games, significantly extending previous results from the literature [19,27,35]. In particular, motivated by the observation that randomized strategies are undesirable or meaningless for certain applications (e.g., medical or product design [23]), we study the setting of DQs under deterministic strategies for both players. Notably, this lead to rather high complexities: Qualitative queries are PSPACE-hard and quantitative reachability is even undecidable.…”
Section: Contributions and Overview In Summary This Paper Makes The Following Contributionsmentioning
confidence: 81%
“…• A comprehensive overview of strategy (Section 3, Table 1) and computational complexity (Section 4, Table 2) of disjunctive reachability-safety queries in stochastic games, significantly extending previous results from the literature [19,27,35]. In particular, motivated by the observation that randomized strategies are undesirable or meaningless for certain applications (e.g., medical or product design [23]), we study the setting of DQs under deterministic strategies for both players. Notably, this lead to rather high complexities: Qualitative queries are PSPACE-hard and quantitative reachability is even undecidable.…”
Section: Contributions and Overview In Summary This Paper Makes The Following Contributionsmentioning
confidence: 81%
“…Executing multi-objective model checking on MDPs for the synthesis of Pareto-optimal policies is an important and non-trivial problem [14]. Despite recent advances [13], [15], [16], [17], [18], existing approaches either use simple iterative methods, or rely on reductions and simplifications to solve the problem using linear programming. This limits their applicability to (i) single-objective problems with multiple strict constraints (for which a single best policy exists); or (ii) unconstrained problems with up to three optimisation objectives.…”
Section: Introductionmentioning
confidence: 99%
“…Multi-objective MDP Various types of objectives known from conventionalsingle-objective-model checking have been lifted to the multi-objective case. These objectives range over ω-regular specifications including LTL [26,27], expected (discounted and non-discounted) total rewards [21,27,28,52,22], stepbounded and reward-bounded reachability probabilities [28,35], and-most relevant for this work-expected long-run average (LRA) rewards [18,11,20], also known as mean pay-offs. For the latter, all current approaches build upon linear programming (LP) which yields a theoretical time-complexity polynomial in the model size.…”
Section: Introductionmentioning
confidence: 99%