Proceedings of the 41st IEEE Conference on Decision and Control, 2002.
DOI: 10.1109/cdc.2002.1184956
|View full text |Cite
|
Sign up to set email alerts
|

Markov decision processes with uncertain transition rates: sensitivity and robust control

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
15
0
1

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 18 publications
(16 citation statements)
references
References 6 publications
0
15
0
1
Order By: Relevance
“…Our work is motivated by the fact that in many practical problems, the transition matrices have to be estimated from data, and this may be a difficult task; see, for example, Kalyanasundaram et al (2001), Feinberg and Shwartz (2002), Abbad and Filar (1992), and Abbad et al (1992). It turns out that estimation errors may have a huge impact on the solution, which is often quite sensitive to changes in the transition probabilities.…”
Section: Introductionmentioning
confidence: 99%
“…Our work is motivated by the fact that in many practical problems, the transition matrices have to be estimated from data, and this may be a difficult task; see, for example, Kalyanasundaram et al (2001), Feinberg and Shwartz (2002), Abbad and Filar (1992), and Abbad et al (1992). It turns out that estimation errors may have a huge impact on the solution, which is often quite sensitive to changes in the transition probabilities.…”
Section: Introductionmentioning
confidence: 99%
“…Inspired by the work of Satia and Lave [35], models involving imprecision have also been applied to the related field of Markov decision processes. Some studies have been devoted to obtain max-min policies, max-max policies and maximal policies when the MDP model is not accurate [5,23].…”
Section: Related Workmentioning
confidence: 99%
“…With those uncertain transition matrices, robust dynamic programming is desired to address the issue of designing approximation method with an appropriate robustness to extend the power of the Bellman Equation. Representative efforts in developing robust dynamic programming can be found in [6]- [10]. One commonly used principle of optimality criterion for robust algorithms is to minimize the maximum value functions for all initial states, which is referred to as robust uniform optimality criterion in this paper.…”
Section: Introductionmentioning
confidence: 99%