Fuzzy Q(λ)-learning algorithm

  • Authors:
  • Roman Zajdel

  • Affiliations:
  • Faculty of Electrical and Computer Engineering, Rzeszow University of Technology, Rzeszow, Poland

  • Venue:
  • ICAISC'10 Proceedings of the 10th international conference on Artificial intelligence and soft computing: Part I
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

The adaptation of temporal differences method TD(λ0) to reinforcement learning algorithms with fuzzy approximation of action-value function is proposed. Eligibility traces are updated using the normalized degree of activation of fuzzy rules. The two types of fuzzy reinforcement learning algorithm are formulated: with discrete and with continuous action values. These new algorithms are practically tested in the control of two typical models of continuous object, like ball-beam and cart-pole system. The achievement results are compared with two popular reinforcement learning algorithms with CMAC and table approximation of action-value function.