Average Continuous Control of Piecewise Deterministic Markov Processes

  • Authors:
  • O. L. V. Costa;F. Dufour

  • Affiliations:
  • oswaldo@lac.usp.br;dufour@math.u-bordeaux1.fr

  • Venue:
  • SIAM Journal on Control and Optimization
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper deals with the long run average continuous control problem of piecewise deterministic Markov processes (PDMPs) taking values in a general Borel space and with compact action space depending on the state variable. The control variable acts on the jump rate and transition measure of the PDMP, and the running and boundary costs are assumed to be positive but not necessarily bounded. Our first main result is to obtain an optimality equation for the long run average cost in terms of a discrete-time optimality equation related to the embedded Markov chain given by the postjump location of the PDMP. Our second main result guarantees the existence of a feedback measurable selector for the discrete-time optimality equation by establishing a connection between this equation and an integro-differential equation. Our final main result is to obtain some sufficient conditions for the existence of a solution for a discrete-time optimality inequality and an ordinary optimal feedback control for the long run average cost using the so-called vanishing discount approach. Two examples are presented illustrating the possible applications of the results developed in the paper.