Describing the emotional states that are expressed in speech
Speech Communication - Special issue on speech and emotion
A corpus-based speech synthesis system with emotion
Speech Communication - Special issue on speech and emotion
The role of voice quality in communicating emotion, mood and attitude
Speech Communication - Special issue on speech and emotion
Vocal communication of emotion: a review of research paradigms
Speech Communication - Special issue on speech and emotion
IEEE Transactions on Audio, Speech, and Language Processing
Generating expressive speech for storytelling applications
IEEE Transactions on Audio, Speech, and Language Processing
Towards Slovak Broadcast News Automatic Recording and Transcribing Service
Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction
Hi-index | 0.00 |
In expressive speech synthesis some method of mimicking the way one specific speaker express emotions is needed. In this work we have studied the suitability of long term prosodic parameters and short term spectral parameters to reflect emotions in speech, by means of the analysis of the results of two automatic emotion classification systems. Those systems have been trained with different emotional monospeaker databases recorded in standard Basque that include six emotions. Both of them are able to differentiate among emotions for a specific speaker with very high identification rates (above 75%), but the models are not applicable to other speakers (identification rates drop to 20%). Therefore in the synthesis process the control of both spectral and prosodic features is essential to get expressive speech and when a change in speaker is desired the values of the parameters should be re-estimated.