A selection-mutation model for q-learning in multi-agent systems

  • Authors:
  • Karl Tuyls;Katja Verbeeck;Tom Lenaerts

  • Affiliations:
  • Computational Modeling Lab, Brussels, Belgium;Computational Modeling Lab, Brussels, Belgium;Computational Modeling Lab, Brussels, Belgium

  • Venue:
  • AAMAS '03 Proceedings of the second international joint conference on Autonomous agents and multiagent systems
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

Although well understood in the single-agent framework, the use of traditional reinforcement learning (RL) algorithms in multi-agent systems (MAS) is not always justified. The feedback an agent experiences in a MAS, is usually influenced by the other agents present in the system. Multi agent environments are therefore non-stationary and convergence and optimality guarantees of RL algorithms are lost. To better understand the dynamics of traditional RL algorithms we analyze the learning process in terms of evolutionary dynamics. More specifically we show how the Replicator Dynamics (RD) can be used as a model for Q-learning in games. The dynamical equations of Q-learning are derived and illustrated by some well chosen experiments. Both reveal an interesting connection between the exploitation-exploration scheme from RL and the selection-mutation mechanisms from evolutionary game theory.