HMM-based emotional speech synthesis using average emotion model

  • Authors:
  • Long Qin;Zhen-Hua Ling;Yi-Jian Wu;Bu-Fan Zhang;Ren-Hua Wang

  • Affiliations:
  • iFLYTEK Speech Lab, University of Science and Technology of China, Hefei;iFLYTEK Speech Lab, University of Science and Technology of China, Hefei;iFLYTEK Speech Lab, University of Science and Technology of China, Hefei;iFLYTEK Speech Lab, University of Science and Technology of China, Hefei;iFLYTEK Speech Lab, University of Science and Technology of China, Hefei

  • Venue:
  • ISCSLP'06 Proceedings of the 5th international conference on Chinese Spoken Language Processing
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents a technique for synthesizing emotional speech based on an emotion-independent model which is called “average emotion” model. The average emotion model is trained using a multi-emotion speech database. Applying a MLLR-based model adaptation method, we can transform the average emotion model to present the target emotion which is not included in the training data. A multi-emotion speech database including four emotions, “neutral”, “happiness”, “sadness”, and “anger”, is used in our experiment. The results of subjective tests show that the average emotion model can effectively synthesize neutral speech and can be adapted to the target emotion model using very limited training data.