A short tutorial on reinforcement learning: review and applications

  • Authors:
  • Chengcheng Li;Larry Pyeatt

  • Affiliations:
  • Computer Science Department, Texas Tech University, Lubbock, Texas;Computer Science Department, Texas Tech University, Lubbock, Texas

  • Venue:
  • Intelligent information processing II
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

Dynamic Programming (DP) has been widely used as an approach solving the Markov Decision Process problem. This paper takes a well-known gambler's problem as an example to compare different DP solutions to the problem, and uses a variety of parameters to explain the results in detail. Ten C++ programs were written to implement the algorithms. The numerical results from gamble's problem and graphical output from the tracking car problem support the conceptual definitions of RL methods.