On multimodal interactive machine translation using speech recognition

  • Authors:
  • Vicent Alabau;Luis Rodríguez-Ruiz;Alberto Sanchis;Pascual Martínez-Gómez;Francisco Casacuberta

  • Affiliations:
  • Universitat Politècnica de València, València, Spain;Universidad de Castilla-La Mancha, Albacete, Spain;Universitat Politècnica de València, València, Spain;University of Tokyo, Tokyo, Japan;Universitat Politècnica de València, València, Spain

  • Venue:
  • ICMI '11 Proceedings of the 13th international conference on multimodal interfaces
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Interactive machine translation (IMT) is an increasingly popular paradigm for semi-automated machine translation, where a human expert is integrated into the core of an automatic machine translation system. The human expert interacts with the IMT system by partially correcting the errors of the system's output. Then, the system proposes a new solution. This process is repeated until the output meets the desired quality. In this scenario, the interaction is typically performed using the keyboard and the mouse. However, speech is also a very interesting input modality since the user does not need to abandon the keyboard to interact with it. In this work, we present a new approach to perform speech interaction in a way that translation and speech inputs are tightly fused. This integration is performed early in the speech recognition step. Thus, the information from the translation models allows the speech recognition system to recover from errors that otherwise would be impossible to amend. In addition, this technique allows to use currently available speech recognition technology. The proposed system achieves an important boost in performance with respect to previous approaches.