Effects of chaotic exploration on reinforcement learning in target capturing task

  • Authors:
  • Koichiro Morihiro;Teijiro Isokawa;Nobuyuki Matsui;Haruhiko Nishimura

  • Affiliations:
  • (Correspd. Tel.: +81 795 44 2294/ Fax: +81 795 44 2045/ E-mail: mori@info.hyogo-u.ac.jp) Hyogo University of Teacher Education, Hyogo 673-1494, Japan and Himeji Institute of Technology, Hyogo 671- ...;Himeji Institute of Technology, Hyogo 671-2201, Japan and Graduate School of Engineering, University of Hyogo, Hyogo 671-2201, Japan;Himeji Institute of Technology, Hyogo 671-2201, Japan and Graduate School of Engineering, University of Hyogo, Hyogo 671-2201, Japan;Graduate School of Applied Informatics, University of Hyogo, Hyogo 650-0044, Japan

  • Venue:
  • International Journal of Knowledge-based and Intelligent Engineering Systems
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

A process of trial and error plays an important role in not only the human learning but also the machine learning. Such a process is called exploration in the reinforcement learning which has originated from experimental studies on learning in psychology. A uniform pseudorandom number generator appears to be suitable for exploration. However, it is known that a chaotic source also provides a random-like sequence as like as a stochastic source. By applying this random-like feature of a deterministic chaotic generator for exploration in a nonstationary shortcut maze problem, we have observed that a deterministic chaotic generator provides a better performance than a stochastic random exploration generator when used for exploration based on a logistic map. In this study, in order to confirm this difference in the performances of the two generators, we examine another nonstationary task - target capturing. The simulation result of this task agrees with the result of our previous study. From the view of multi-agent system, it is an inhomogeneous or heterogeneous system composed of some kinds of agents in many cases. In such situations, the exploration of them is not uniform. Chaotic exploration may suit well this heterogeneity in such a multi-agent system.