Fast reinforcement learning of dialogue policies using stable function approximation

  • Authors:
  • Matthias Denecke;Kohji Dohsaka;Mikio Nakano

  • Affiliations:
  • Communication Science Laboratories, Nippon Telegraph and Telephone Corporation, Atsugi Kanagawa;Communication Science Laboratories, Nippon Telegraph and Telephone Corporation, Atsugi Kanagawa;Communication Science Laboratories, Nippon Telegraph and Telephone Corporation, Atsugi Kanagawa

  • Venue:
  • IJCNLP'04 Proceedings of the First international joint conference on Natural Language Processing
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

We propose a method to speed up reinforcement learning of policies for spoken dialogue systems. This is achieved by combining a coarse grained abstract representation of states and actions with learning only in frequently visited states. The value of unsampled states is approximated by a linear interpolation of known states. Experiments show that the proposed method effectively optimizes dialogue strategies for frequently visited dialogue states.