The ATR multilingual speech-to-speech translation system

  • Authors:
  • S. Nakamura;K. Markov;H. Nakaiwa;G. Kikui;H. Kawai;T. Jitsuhiro;Jin-Song Zhang;H. Yamamoto;E. Sumita;S. Yamamoto

  • Affiliations:
  • ATR Spoken Language Translation Res. Labs., Kyoto, Japan;-;-;-;-;-;-;-;-;-

  • Venue:
  • IEEE Transactions on Audio, Speech, and Language Processing
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we describe the ATR multilingual speech-to-speech translation (S2ST) system, which is mainly focused on translation between English and Asian languages (Japanese and Chinese). There are three main modules of our S2ST system: large-vocabulary continuous speech recognition, machine text-to-text (T2T) translation, and text-to-speech synthesis. All of them are multilingual and are designed using state-of-the-art technologies developed at ATR. A corpus-based statistical machine learning framework forms the basis of our system design. We use a parallel multilingual database consisting of over 600 000 sentences that cover a broad range of travel-related conversations. Recent evaluation of the overall system showed that speech-to-speech translation quality is high, being at the level of a person having a Test of English for International Communication (TOEIC) score of 750 out of the perfect score of 990.