A consideration of human immunity-based reinforcement learning with continuous states

  • Authors:
  • Shu Hosokawa;Kazushi Nakano;Kazunori Sakurama

  • Affiliations:
  • Department of Electronic Engineering, The University of Electro-Communications, Chofu, Tokyo, Japan 182-8585;Department of Electronic Engineering, The University of Electro-Communications, Chofu, Tokyo, Japan 182-8585;Department of Electronic Engineering, The University of Electro-Communications, Chofu, Tokyo, Japan 182-8585

  • Venue:
  • Artificial Life and Robotics
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

Many reinforcement learning methods have been studied on the assumption that a state is discretized and the environment size is predetermined. However, an operating environment may have a continuous state and its size may not be known in advance, e.g., in robot navigation and control. When applying these methods to the environment described above, we may need a large amount of time for learning or failing to learn. In this study, we improve our previous human immunity-based reinforcement learning method so that it will work in continuous state space environments. Since our method selects an action based on the distance between the present state and the memorized action, information about the environment (e.g., environment size) is not required in advance. The validity of our method is demonstrated through simulations for the swingup control of an inverted pendulum.