Policy transfer via Markov logic networks

  • Authors:
  • Lisa Torrey;Jude Shavlik

  • Affiliations:
  • University of Wisconsin, Madison, WI;University of Wisconsin, Madison, WI

  • Venue:
  • ILP'09 Proceedings of the 19th international conference on Inductive logic programming
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

We propose using a statistical-relational model, the Markov Logic Network, for knowledge transfer in reinforcement learning. Our goal is to extract relational knowledge from a source task and use it to speed up learning in a related target task. We show that Markov Logic Networks are effective models for capturing both source-task Q-functions and source-task policies. We apply them via demonstration, which involves using them for decision making in an initial stage of the target task before continuing to learn. Through experiments in the RoboCup simulated-soccer domain, we show that transfer via Markov Logic Networks can significantly improve early performance in complex tasks, and that transferring policies is more effective than transferring Q-functions.