Spoken Dialog Challenge 2010: comparison of live and control test results

  • Authors:
  • Alan W. Black;Susanne Burger;Alistair Conkie;Helen Hastie;Simon Keizer;Oliver Lemon;Nicolas Merigaud;Gabriel Parent;Gabriel Schubiner;Blaise Thomson;Jason D. Williams;Kai Yu;Steve Young;Maxine Eskenazi

  • Affiliations:
  • Carnegie Mellon University, Pittsburgh;Carnegie Mellon University, Pittsburgh;AT&T Labs -- Research, Florham Park, NJ;Heriot-Watt University, Edinburgh, UK;Cambridge University, Cambridge, UK;Heriot-Watt University, Edinburgh, UK;Heriot-Watt University, Edinburgh, UK;Carnegie Mellon University, Pittsburgh;Carnegie Mellon University, Pittsburgh;Cambridge University, Cambridge, UK;AT&T Labs -- Research, Florham Park, NJ;Cambridge University, Cambridge, UK;Cambridge University, Cambridge, UK;Carnegie Mellon University, Pittsburgh

  • Venue:
  • SIGDIAL '11 Proceedings of the SIGDIAL 2011 Conference
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

The Spoken Dialog Challenge 2010 was an exercise to investigate how different spoken dialog systems perform on the same task. The existing Let's Go Pittsburgh Bus Information System was used as a task and four teams provided systems that were first tested in controlled conditions with speech researchers as users. The three most stable systems were then deployed to real callers. This paper presents the results of the live tests, and compares them with the control test results. Results show considerable variation both between systems and between the control and live tests. Interestingly, relatively high task completion for controlled tests did not always predict relatively high task completion for live tests. Moreover, even though the systems were quite different in their designs, we saw very similar correlations between word error rate and task completion for all the systems. The dialog data collected is available to the research community.