Mathematics of Operations Research
Denumerable semi-Markov decision chains with small interest rates
Annals of Operations Research
Mathematics of Operations Research
Discrete-time controlled Markov processes with average cost criterion: a survey
SIAM Journal on Control and Optimization
On the relation between recurrence and ergodicity properties in denumerable Markov decision chains
Mathematics of Operations Research
Mathematics of Operations Research
M/G/1-Type Markov Processes: A Tutorial
Performance Evaluation of Complex Systems: Techniques and Tools, Performance 2002, Tutorial Lectures
THE DEVIATION MATRIX OF A CONTINUOUS-TIME MARKOV CHAIN
Probability in the Engineering and Informational Sciences
Optimal Control of Ergodic Continuous-Time Markov Chains with Average Sample-Path Rewards
SIAM Journal on Control and Optimization
Stochastic Learning and Optimization: A Sensitivity-Based Approach (International Series on Discrete Event Dynamic Systems)
Discrete-Event Control of Stochastic Networks: Multimodularity and Regularity (Lecture Notes in Mathematics)
Continuous-Time Markov Decision Processes with Discounted Rewards: The Case of Polish Spaces
Mathematics of Operations Research
Series Expansions For Finite-State Markov Chains
Probability in the Engineering and Informational Sciences
Brief paper: Policy iteration based feedback control
Automatica (Journal of IFAC)
Series Expansions for Continuous-Time Markov Processes
Operations Research
Computers & Mathematics with Applications
A functional approximation for the M/G/1/N queue
Discrete Event Dynamic Systems
Hi-index | 0.00 |
We present an update formula that allows the expression of the deviation matrix of a continuous-time Markov process with denumerable state space having generator matrix Q* through a continuous-time Markov process with generator matrix Q. We show that under suitable stability conditions the algorithm converges at a geometric rate. By applying the concept to three different examples, namely, the M/M/1 queue with vacations, the M/G/1 queue, and a tandem network, we illustrate the broad applicability of our approach. For a problem in admission control, we apply our approximation algorithm to Markov decision theory for computing the optimal control policy. Numerical examples are presented to highlight the efficiency of the proposed algorithm.