A dynamic programming strategy to balance exploration and exploitation in the bandit problem

  • Authors:
  • Olivier Caelen;Gianluca Bontempi

  • Affiliations:
  • Computer Science Department, Université Libre de Bruxelles, Bruxelles, Belgium 1050;Computer Science Department, Université Libre de Bruxelles, Bruxelles, Belgium 1050

  • Venue:
  • Annals of Mathematics and Artificial Intelligence
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

The K-armed bandit problem is a well-known formalization of the exploration versus exploitation dilemma. In this learning problem, a player is confronted to a gambling machine with K arms where each arm is associated to an unknown gain distribution. The goal of the player is to maximize the sum of the rewards. Several approaches have been proposed in literature to deal with the K-armed bandit problem. This paper introduces first the concept of "expected reward of greedy actions" which is based on the notion of probability of correct selection (PCS), well-known in simulation literature. This concept is then used in an original semi-uniform algorithm which relies on the dynamic programming framework and on estimation techniques to optimally balance exploration and exploitation. Experiments with a set of simulated and realistic bandit problems show that the new DP-greedy algorithm is competitive with state-of-the-art semi-uniform techniques.