Transfer of task representation in reinforcement learning using policy-based proto-value functions

  • Authors:
  • Eliseo Ferrante;Alessandro Lazaric;Marcello Restelli

  • Affiliations:
  • Politecnico di Milano, piazza Leonardo Da Vinci, Milan, Italy;Politecnico di Milano, piazza Leonardo Da Vinci, Milan, Italy;Politecnico di Milano, piazza Leonardo Da Vinci, Milan, Italy

  • Venue:
  • Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems - Volume 3
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

Reinforcement Learning research is traditionally devoted to solve single-task problems. Therefore, anytime a new task is faced, learning must be restarted from scratch. Recently, several studies have addressed the issue of reusing the knowledge acquired in solving previous related tasks by transferring information about policies and value functions. In this paper, we analyze the use of proto-value functions under the transfer learning perspective. Proto-value functions are effective basis functions for the approximation of value functions defined over the graph obtained by a random walk on the environment. The definition of this graph is a key aspect in transfer transfer problems in which both the reward function and the dynamics change. Therefore, we introduce policy-based proto-value functions, which can be obtained by considering the graph generated by a random walk guided by the optimal policy of one of the tasks at hand. We compare the effectiveness of policy-based and standard proto-value functions, on different transfer problems defined on a simple grid-world environment.