2012 IEEE 51st IEEE Conference on Decision and Control (CDC) 2012
DOI: 10.1109/cdc.2012.6426174
|View full text |Cite
|
Sign up to set email alerts
|

Robust control of uncertain Markov Decision Processes with temporal logic specifications

Abstract: Abstract-We present a method for designing robust controllers for dynamical systems with linear temporal logic specifications. We abstract the original system by a finite Markov Decision Process (MDP) that has transition probabilities in a specified uncertainty set. A robust control policy for the MDP is generated that maximizes the worst-case probability of satisfying the specification over all transition probabilities in the uncertainty set. To do this, we use a procedure from probabilistic model checking to… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
121
0
1

Year Published

2013
2013
2022
2022

Publication Types

Select...
8

Relationship

1
7

Authors

Journals

citations
Cited by 123 publications
(127 citation statements)
references
References 28 publications
0
121
0
1
Order By: Relevance
“…There has been significant recent work using Markov decision processes, temporal logic specifications and model checking for generating controllers of autonomous systems. These works differ in the temporal logic specifications used and include approaches using the branching time logic PCTL [36,22], linear time temporal logic LTL [35,7], metric temporal logic [11], rewards [30] and multi-objective queries [23,21]. Also, both partially observable Markov decision processes, e.g.…”
Section: Introductionmentioning
confidence: 99%
“…There has been significant recent work using Markov decision processes, temporal logic specifications and model checking for generating controllers of autonomous systems. These works differ in the temporal logic specifications used and include approaches using the branching time logic PCTL [36,22], linear time temporal logic LTL [35,7], metric temporal logic [11], rewards [30] and multi-objective queries [23,21]. Also, both partially observable Markov decision processes, e.g.…”
Section: Introductionmentioning
confidence: 99%
“…The proof follows closely the ones in [B.1] (Vol. II, Section 2.4) and [20]. Those proofs refer to a control setting, where the optimal action (control) can be selected.…”
Section: A4 Entropy Modelmentioning
confidence: 99%
“…Conversely, in the verification setting, the contraction needs to hold across all available actions, because we consider the worst case resolution of nondeterminism (universal quantification). Further, as in [20], we quantify across all nature behaviors: this is possible due to Assumption 2.2. For the sake of brevity, in the following we will only consider the calculation of P r min s , but the same reasoning applies also for the maximization problem.…”
Section: A4 Entropy Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…These works are incomplete, however, in that methods to construct the approximating mdp for the robot are not presented; only planning over an existing abstraction is discussed. Uncertain mdps, where transition probabilities can belong to sets of values, have also been employed to provide a hierarchical abstraction and improve computational complexity [22]. Strong assumptions must be made on the structure of this abstraction, many of which are difficult to realize for physical systems.…”
Section: Related Workmentioning
confidence: 99%