2006
DOI: 10.1561/0900000002
|View full text |Cite
|
Sign up to set email alerts
|

Monotonicity in Markov Reward and Decision Chains: Theory and Applications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
51
0

Year Published

2010
2010
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 73 publications
(51 citation statements)
references
References 54 publications
(94 reference statements)
0
51
0
Order By: Relevance
“…A standard approach for studying the optimal policies of MDPs is to explore the first-and/or second-order properties of the optimal cost function (see Koole 2006). Optimal cost functions for multivariate MDPs (like ours) are typically shown to be convex in each dimension of the state space.…”
Section: Introductionmentioning
confidence: 99%
“…A standard approach for studying the optimal policies of MDPs is to explore the first-and/or second-order properties of the optimal cost function (see Koole 2006). Optimal cost functions for multivariate MDPs (like ours) are typically shown to be convex in each dimension of the state space.…”
Section: Introductionmentioning
confidence: 99%
“…For both models, we show via dynamic programming that (i) the optimal allocation policy has a work-conservation property that implies when the system is not empty, the optimal policy is not allowed to keep all computing resources idle, (ii) the optimal number of servers follows a step function with as extreme policy the bang-bang control policy, which means a facility receives all computing resources or none at all, and moreover (iii) we also provide the conditions under which the bang-bang control policy is optimal. The techniques to prove such results are based on monotonicity properties of the dynamic programming relative value function (see, e.g., Koole 1998Koole , 2006Rykov 2001).…”
Section: Introductionmentioning
confidence: 99%
“…To analyze the value function V n (x, y) in §4.4.1.3, we employ the event based dynamic programming approach introduced by Koole (1998Koole ( , 2006. To this end, let V denote the set of all functions v : S → R and let f, f 1 , ..., f m+2 ∈ V. We define the following operators…”
Section: Mdp Formulation With Bounded Transition Ratesmentioning
confidence: 99%
“…These operators are variations to operators defined by Koole (1998Koole ( , 2004Koole ( , 2006 and are originally intended to model various common queueing mechanisms such as arrival control (T AC(i) ), transfer departures from multi-server tandem queues (T TD(i) ), and departures from multi-server queues (T D(i) ), while the operators T cost f (x, y), T env and T unif are mainly convenient for bookkeeping. The Bellman recursion for our MDP, (4.4), can now be written succinctly as…”
Section: Mdp Formulation With Bounded Transition Ratesmentioning
confidence: 99%
See 1 more Smart Citation