Optimising incremental generation for spoken dialogue systems: reducing the need for fillers

  • Authors:
  • Nina Dethlefs;Helen Hastie;Verena Rieser;Oliver Lemon

  • Affiliations:
  • Heriot Watt University, Edinburgh;Heriot Watt University, Edinburgh;Heriot Watt University, Edinburgh;Heriot Watt University, Edinburgh

  • Venue:
  • INLG '12 Proceedings of the Seventh International Natural Language Generation Conference
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

Recent studies have shown that incremental systems are perceived as more reactive, natural, and easier to use than non-incremental systems. However, previous work on incremental NLG has not employed recent advances in statistical optimisation using machine learning. This paper combines the two approaches, showing how the update, revoke and purge operations typically used in incremental approaches can be implemented as state transitions in a Markov Decision Process. We design a model of incremental NLG that generates output based on micro-turn interpretations of the user's utterances and is able to optimise its decisions using statistical machine learning. We present a proof-of-concept study in the domain of Information Presentation (IP), where a learning agent faces the trade-off of whether to present information as soon as it is available (for high reactiveness) or else to wait until input ASR hypotheses are more reliable. Results show that the agent learns to avoid long waiting times, fillers and self-corrections, by re-ordering content based on its confidence.