Improving Coordination with Communication in Multi-Agent Reinforcement Learning

  • Authors:
  • Daniel Szer;Francois Charpillet

  • Affiliations:
  • INRIA-LORIA;INRIA-LORIA

  • Venue:
  • ICTAI '04 Proceedings of the 16th IEEE International Conference on Tools with Artificial Intelligence
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

In the following paper we present a new algorithm for cooperative reinforcement learning in multi-agent systems. We consider autonomous and independently learning agents, and we seek to obtain an optimal solution for the team as a whole while keeping the learning as much decentralized as possible. Coordination between agents occurs through communication, namely the mutual notification algorithm. We define the learning problem as a decentralized process using the MDP formalism. We then give an optimality criterion and prove the convergence of the algorithm for deterministic environments.We introduce variable and hierarchical communication strategies which considerably reduce the number of communications. Finally we study the convergence properties and communication over-head on a small example.