On the performance of on-line concurrent reinforcement learners

  • Authors:
  • Bikramjit Banerjee

  • Affiliations:
  • Tulane University, New Orleans, LA

  • Venue:
  • Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

Multiagent Learning (MAL) is significantly complicated relative to Machine Learning (ML) by the fact that multiple learners render each other's environments non-stationary. While ML focuses on learning a fixed target function, MAL deals with learning a "moving target function". In contrast to classical Reinforcement Learning, MAL deals with an extra level of uncertainty in the form of the behaviors of the other learners in the domain. Existing learning methods provide guarantees about the performance of the learners only in the limit since a learner approaches its desired behavior asymptotically. There is little insight into how well or how poorly an on-line learner can perform while it is learning. This is the core problem studied in this thesis, resulting in the following contributions.