2000
DOI: 10.1006/jmaa.2000.6819
|View full text |Cite
|
Sign up to set email alerts
|

Multiple Objective Nonatomic Markov Decision Processes with Total Reward Criteria

Abstract: We consider a Markov decision process with an uncountable state space and multiple rewards. For each policy, its performance is evaluated by a vector of total expected rewards. Under the standard continuity assumptions and the additional assumption that all initial and transition probabilities are nonatomic, we prove that the set of performance vectors for all policies is equal to the set of performance Ž . vectors for nonrandomized Markov policies. This result implies the existence of Ž . optimal nonrandomize… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
11
0

Year Published

2002
2002
2020
2020

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(11 citation statements)
references
References 22 publications
0
11
0
Order By: Relevance
“…It seems that establishing this characterization result could be quite involving, especially for general CTMDP models in Borel spaces. Instead, like in [12,13] and [36] for discrete-time and continuous-time problems with total undiscounted and discounted criteria and [31] focusing on the performance analysis of queueing networks, we pass the average constrained CTMDP problem from the infinite dimensional framework (in the space of measures) to the finite dimensional framework by investigating the space of performance vectors.…”
Section: Introductionmentioning
confidence: 99%
“…It seems that establishing this characterization result could be quite involving, especially for general CTMDP models in Borel spaces. Instead, like in [12,13] and [36] for discrete-time and continuous-time problems with total undiscounted and discounted criteria and [31] focusing on the performance analysis of queueing networks, we pass the average constrained CTMDP problem from the infinite dimensional framework (in the space of measures) to the finite dimensional framework by investigating the space of performance vectors.…”
Section: Introductionmentioning
confidence: 99%
“…This result was established in [7] as a corollary of the following fact: the performance set for nonrandomized Markov policies coincides with Research of this coauthor was partially supported by NSF Grant DMI-9908258 with the performance set for all policies in a nonatomic MDP satisfying continuity and compactness conditions. Continuity and compactness conditions were essential for the proofs in Feinberg and Piunovskiy [7].…”
Section: Introductionmentioning
confidence: 81%
“…Lemma 9 in Feinberg and Piunovskiy [7] implies that there is a poligy y in the new MDP such that fi"(P7) = Rn(P") and RN+"(p7) = Rn (Pn, T ) Consider a sequence Ek \ 0. We define TI > 0 such that for all n = 1 , .…”
Section: Lemma 3 For Any Policy W Andmentioning
confidence: 99%
See 2 more Smart Citations