A multi-agent reinforcement learning method for a partially-observable competitive game

  • Authors:
  • Yoichiro Matsuno;Tatsuya Ymazaki;Shin Ishii;Jun Matsuno

  • Affiliations:
  • Osaka Gakuin University;-;-;-

  • Venue:
  • Proceedings of the fifth international conference on Autonomous agents
  • Year:
  • 2001

Quantified Score

Hi-index 0.00

Visualization

Abstract

This article proposes a reinforcement learning (RL) method based on an actor-critic architecture, which can be applied to partially-observable multi-agent competitive games. As an example, we deal with a card game “Hearts”. In our method, the actor plays so as to enlarge the expected temporal-difference error, which is obtained based on the estimation of the state transition. The state transition is estimated by taking the inferred card distribution and the other player's action models into account.