Reinforcement learning for mapping instructions to actions

  • Authors:
  • S. R. K. Branavan;Harr Chen;Luke S. Zettlemoyer;Regina Barzilay

  • Affiliations:
  • Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology;Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology;Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology;Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology

  • Venue:
  • ACL '09 Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP: Volume 1 - Volume 1
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we present a reinforcement learning approach for mapping natural language instructions to sequences of executable actions. We assume access to a reward function that defines the quality of the executed actions. During training, the learner repeatedly constructs action sequences for a set of documents, executes those actions, and observes the resulting reward. We use a policy gradient algorithm to estimate the parameters of a log-linear model for action selection. We apply our method to interpret instructions in two domains --- Windows troubleshooting guides and game tutorials. Our results demonstrate that this method can rival supervised learning techniques while requiring few or no annotated training examples.