Impacts of machine translation and speech synthesis on speech-to-speech translation

  • Authors:
  • Kei Hashimoto;Junichi Yamagishi;William Byrne;Simon King;Keiichi Tokuda

  • Affiliations:
  • Nagoya Institute of Technology, Department of Computer Science and Engineering, Nagoya, Japan;University of Edinburgh, Centre for Speech Technology Research, Edinburgh, United Kingdom;Cambridge University, Engineering Department, Cambridge, United Kingdom;University of Edinburgh, Centre for Speech Technology Research, Edinburgh, United Kingdom;Nagoya Institute of Technology, Department of Computer Science and Engineering, Nagoya, Japan

  • Venue:
  • Speech Communication
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper analyzes the impacts of machine translation and speech synthesis on speech-to-speech translation systems. A typical speech-to-speech translation system consists of three components: speech recognition, machine translation and speech synthesis. Many techniques have been proposed for integration of speech recognition and machine translation. However, corresponding techniques have not yet been considered for speech synthesis. The focus of the current work is machine translation and speech synthesis, and we present a subjective evaluation designed to analyze their impact on speech-to-speech translation. The results of these analyses show that the naturalness and intelligibility of the synthesized speech are strongly affected by the fluency of the translated sentences. In addition, several features were found to correlate well with the average fluency of the translated sentences and the average naturalness of the synthesized speech.