Multiple Reward Criterion for Cooperative Behavior Acquisition in a Muliagent Environment

  • Authors:
  • Eiji Uchibe;Minoru Asada

  • Affiliations:
  • -;-

  • Venue:
  • RoboCup-99: Robot Soccer World Cup III
  • Year:
  • 2000

Quantified Score

Hi-index 0.00

Visualization

Abstract

An extended value function is discussed in the context of multiple behavior coordination, especially in a dynamically changing multiagent environment. Unlike the traditional weighted sum of several reward functions, we define a vectorized value function which evaluates the current action strategy by introducing a discounted matrix to integrate several reward functions. Owing to the extension of the value function, the learning robot can estimate the future multiple rewards from the environment appropriately not suffering from the weighting problem. The proposed method is applied to a simplified soccer game. Computer simulations are shown and a discussion is given.