Modeling Emotion and Attitude in Speech by Means of Perceptually Based Parameter Values

  • Authors:
  • Sylvie J. L. Mozziconacci

  • Affiliations:
  • IPO, Center for User-System Interaction, Eindhoven, The Netherlands

  • Venue:
  • User Modeling and User-Adapted Interaction
  • Year:
  • 2001

Quantified Score

Hi-index 0.00

Visualization

Abstract

This study focuses on the perception of emotion and attitude in speech. The ability to identify vocal expressions of emotion and/or attitude in speech material was investigated. Systematic perception experiments were carried out to determine optimal values for the acoustic parameters: pitch level, pitch range and speech rate. Speech was manipulated by varying these parameters around the values found in a selected subset of the speech material which consisted of two sentences spoken by a male speaker expressing seven emotions or attitudes: neutrality, joy, boredom, anger, sadness, fear, and indignation. Listening tests were carried out with this speech material, and optimal values for pitch level, pitch range, and speech rate were derived for the generation of speech expressing emotion or attitude, from a neutral utterance. These values were perceptually tested in re-synthesized speech and in synthetic speech generated from LPC-coded diphones.