Discounted Continuous-Time Markov Decision Processes with Constraints: Unbounded Transition and Loss Rates

  • Authors:
  • Xianping Guo;Alexei Piunovskiy

  • Affiliations:
  • School of Mathematics and Computational Science, Zhongshan University, 510275 Guangzhou, People's Republic of China;Department of Mathematical Sciences, The University of Liverpool, Liverpool L69 7ZL, United Kingdom

  • Venue:
  • Mathematics of Operations Research
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper deals with denumerable continuous-time Markov decision processes (MDP) with constraints. The optimality criterion to be minimized is expected discounted loss, while several constraints of the same type are imposed. The transition rates may be unbounded, the loss rates are allowed to be unbounded as well (from above and from below), and the policies may be history-dependent and randomized. Based on Kolmogorov's forward equation and Dynkin's formula, we remind the reader about the Bellman equation, introduce and study occupation measures, reformulate the optimization problem as a (primary) linear program, provide the form of optimal policies for a constrained optimization problem here, and establish the duality between the convex analytic approach and dynamic programming. Finally, a series of examples is given to illustrate all of our results.