Exploration strategies for learned probabilities in smart terrain

  • Authors:
  • John Sullins

  • Affiliations:
  • Youngstown State University, Youngstown, Ohio

  • Venue:
  • MLDM'11 Proceedings of the 7th international conference on Machine learning and data mining in pattern recognition
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Consider a mobile agent (such as a robot) surrounded by objects that may or may not meet its needs. An important goal of such an agent is to learn probabilities that different types of objects meet needs, based on objects it has previously explored. This requires a rational strategy for determining which objects to explore next based on distances to objects, prevalence of similar objects, and amount of information the agent expects to gain. We define information gain in terms of how additional examples increase the certainty of the probabilities (represented as beta distributions), based on how that certainty reduces future travel time by preventing the agent from moving to objects which do not actually meet needs. This is used to create a smart terrain-based influence map in which objects send signals proportional to their information gain (with inverse falloff over distance) to enable simple agent navigation to those objects.