Active policy iteration: efficient exploration through active learning for value function approximation in reinforcement learning

  • Authors:
  • Takayuki Akiyama;Hirotaka Hachiya;Masashi Sugiyama

  • Affiliations:
  • Department of Computer Science, Tokyo Institute of Technology;Department of Computer Science, Tokyo Institute of Technology;Department of Computer Science, Tokyo Institute of Technology

  • Venue:
  • IJCAI'09 Proceedings of the 21st international jont conference on Artifical intelligence
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Appropriately designing sampling policies is highly important for obtaining better control policies in reinforcement learning. In this paper, we first show that the least-squares policy iteration (LSPI) framework allows us to employ statistical active learning methods for linear regression. Then we propose a design method of good sampling policies for efficient exploration, which is particularly useful when the sampling cost of immediate rewards is high. We demonstrate the usefulness of the proposed method, named active policy iteration (API), through simulations with a batting robot.