Personalising speech-to-speech translation in the EMIME project

  • Authors:
  • Mikko Kurimo;William Byrne;John Dines;Philip N. Garner;Matthew Gibson;Yong Guan;Teemu Hirsimäki;Reima Karhila;Simon King;Hui Liang;Keiichiro Oura;Lakshmi Saheer;Matt Shannon;Sayaka Shiota;Jilei Tian;Keiichi Tokuda;Mirjam Wester;Yi-Jian Wu;Junichi Yamagishi

  • Affiliations:
  • Aalto University, Finland;University of Cambridge, UK;Idiap Research Institute, Switzerland;Idiap Research Institute, Switzerland;University of Cambridge, UK;Nokia Research Center, Beijing, China;Aalto University, Finland;Aalto University, Finland;University of Edinburgh, UK;Idiap Research Institute, Switzerland;Nagoya Institute of Technology, Japan;Idiap Research Institute, Switzerland;University of Cambridge, UK;Nagoya Institute of Technology, Japan;Nokia Research Center, Beijing, China;Nagoya Institute of Technology, Japan;University of Edinburgh, UK;Nagoya Institute of Technology, Japan;University of Edinburgh, UK

  • Venue:
  • ACLDemos '10 Proceedings of the ACL 2010 System Demonstrations
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

In the EMIME project we have studied unsupervised cross-lingual speaker adaptation. We have employed an HMM statistical framework for both speech recognition and synthesis which provides transformation mechanisms to adapt the synthesized voice in TTS (text-to-speech) using the recognized voice in ASR (automatic speech recognition). An important application for this research is personalised speech-to-speech translation that will use the voice of the speaker in the input language to utter the translated sentences in the output language. In mobile environments this enhances the users' interaction across language barriers by making the output speech sound more like the original speaker's way of speaking, even if she or he could not speak the output language.