An online multi-agent co-operative learning algorithm in POMDPs

  • Authors:
  • Fei Liu;Guangzhou Zeng

  • Affiliations:
  • School of Computer Science and Technology, Shandong University, Jinan, People's Republic of China;School of Computer Science and Technology, Shandong University, Jinan, People's Republic of China

  • Venue:
  • Journal of Experimental & Theoretical Artificial Intelligence
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

Solving partially observable Markov decision processes (POMDPs) is a complex task that is often intractable. This paper examines the problem of finding an optimal policy for POMDPs. While a lot of effort has been made to develop algorithms to solve POMDPs, the question of automatically finding good low-dimensional spaces in multi-agent co-operative learning domains has not been explored thoroughly. To identify this question, an online algorithm CMEAS is presented to improve the POMDP model. This algorithm is based on a look-ahead search to find the best action to execute at each cycle. Thus the overwhelming complexity of computing a policy for each possible situation is avoided. A series of simulations demonstrate this good strategy and performance of the proposed algorithm when multiple agents co-operate to find an optimal policy for POMDPs.