Stochastic policy search for variance-penalized semi-Markov control

  • Authors:
  • Abhijit Gosavi;Mandar Purohit

  • Affiliations:
  • Missouri University of Science and Technology, Rolla, MO;ESRI, Redlands, CA

  • Venue:
  • Proceedings of the Winter Simulation Conference
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

The variance-penalized metric in Markov decision processes (MDPs) seeks to maximize the average reward minus a scalar times the variance of rewards. In this paper, our goal is to study the same metric in the context of the semi-Markov decision process (SMDP). In the SMDP, unlike the MDP, the time spent in each transition is not identical and may in fact be a random variable. We first develop an expression for the variance of rewards in the SMDPs, and then formulate the VP-SMDP. Our interest here is in solving the problem without generating the underlying transition probabilities of the Markov chains. We propose the use of two stochastic search techniques, namely simultaneous perturbation and learning automata, to solve the problem; these techniques use stochastic policies and can be used within simulators, thereby avoiding the generation of the transition probabilities.