Learning to approach a moving ball with a simulated two-wheeled robot

  • Authors:
  • Felix Flentge

  • Affiliations:
  • Department of Computer Science, Johannes Gutenberg-University Mainz, Mainz, Germany

  • Venue:
  • RoboCup 2005
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

We show how a two-wheeled robot can learn to approach a moving ball using Reinforcement Learning. The robot is controlled by setting the velocities of its two wheels. It has to reach the ball under certain conditions to be able to kick it towards a given target. In order to kick, the ball has to be in front of the robot. The robot also has to reach the ball at a certain angle in relation to the target, because the ball is always kicked in the direction from the center of the robot to the ball. The robot learns which velocity differences should be applied to the wheels: one of the wheels is set to the maximum velocity, the other one according to this difference. We apply a REINFORCE algorithm [1] in combination with some kind of extended Growing Neural Gas (GNG) [2] to learn these continuous actions. The resulting algorithm, called ReinforceGNG, is tested in a simulated environment with and without noise.