Applying the policy gradient method to behavior learning in multiagent systems: The pursuit problem

  • Authors:
  • Seiji Ishihara;Harukazu Igarashi

  • Affiliations:
  • School of Engineering, Kinki University, Higashi-Hiroshima, 739-2116 Japan;School of Engineering, Kinki University, Higashi-Hiroshima, 739-2116 Japan

  • Venue:
  • Systems and Computers in Japan
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

In the field of multiagent systems, some methods use the policy gradient method for behavior learning. In these methods, the learning problem in the multiagent system is reduced to each agent's independent learning problem by adopting an autonomous distributed behavior determination method. That is, a probabilistic policy that contains parameters is used as the policy of each agent, and the parameters are updated while calculating the maximum gradient so as to maximize the expectation value of the reward. In this paper, first, recognizing the action determination problem at each time step to be a minimization problem for some objective function, the Boltzmann distribution, in which this objective function is the energy function, was adopted as the probabilistic policy. Next, we showed that this objective function can be expressed by such terms as the value of the state, the state action rule, and the potential. Further, as a result of an experiment applying this method to a pursuit problem, good policy was obtained and this method was found to be flexible so that it can be adapted to use of heuristics and to modification of behavioral constraint and objective in the policy. © 2006 Wiley Periodicals, Inc. Syst Comp Jpn, 37(10): 101–109, 2006; Published online in Wiley InterScience (). DOI 10.1002/scj.20248