Can we hear the prosody of smile?

  • Authors:
  • Véronique Aubergé;Marie Cathiard

  • Affiliations:
  • Institut de la Communication Parlée, Université Stendhal/INPG, Du, 1180 Av Centrale, BP25-38040 Grenoble Cedex 9, France;Institut de la Communication Parlée, Université Stendhal/INPG, Du, 1180 Av Centrale, BP25-38040 Grenoble Cedex 9, France

  • Venue:
  • Speech Communication - Special issue on speech and emotion
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

Visual expression alone (a smile or a laugh) is often enough to identify an emotion such as amusement: in a perception task, subjects correctly identified visual and audio-visual stimuli of amused speech to the same degree (Proceedings of ICSLP, Sydney, 1998, p. 559), but even from the acoustic signal alone it has been demonstrated that the consequences of a (mechanical) smile gesture can be perceived as amusement (JASA 96 (1994) 2101). A hypothesis developed in the present work is that the expression of amusement in speech involves specific control of prosody and cannot be reduced simply to a change in voice quality as a consequence of the facial smile gesture. Speech stimuli were produced by French speakers for various tasks (spontaneous amusement, simulated amusement, mechanical smiling, ...), and in a first experiment, listeners were able to identify speech from the spontaneous smile condition as more amused than the "mechanical smile". It was shown from a second experiment that, even under clear visual conditions, the auditory modality contributes to audio-visual perception. A McGurk paradigm applied to discordant amused/mechanical stimuli clearly showed that acoustic information interacts with the visual decoding. The stimuli were analysed using a set of parameters chosen following Tartter (Percept. Psychophys. 27 (1980) 24), Banse and Scherer (J. Pers. Soc. Psych. 170 (1996) 614) and Mozziconacci (PhD Thesis, Eindhoren University, 1998). The prosodic parameters affected in the expression of amusement are primarily intensity and F0 declination, but they are different for different speakers. Our results confirm the finding by Mozziconacci (PhD Thesis, Eindhoren University, 1998), that there may be numerous ways of using the same parameters to express emotions such as amusement..