Recognising realistic emotions and affect in speech: State of the art and lessons learnt from the first challenge

  • Authors:
  • Björn Schuller;Anton Batliner;Stefan Steidl;Dino Seppi

  • Affiliations:
  • Institute for Human-Machine Communication, Technische Universität München, Germany;Pattern Recognition Lab, University of Erlangen-Nuremberg, Germany;Pattern Recognition Lab, University of Erlangen-Nuremberg, Germany;ESAT, Katholieke Universiteit Leuven, Belgium

  • Venue:
  • Speech Communication
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

More than a decade has passed since research on automatic recognition of emotion from speech has become a new field of research in line with its 'big brothers' speech and speaker recognition. This article attempts to provide a short overview on where we are today, how we got there and what this can reveal us on where to go next and how we could arrive there. In a first part, we address the basic phenomenon reflecting the last fifteen years, commenting on databases, modelling and annotation, the unit of analysis and prototypicality. We then shift to automatic processing including discussions on features, classification, robustness, evaluation, and implementation and system integration. From there we go to the first comparative challenge on emotion recognition from speech - the INTERSPEECH 2009 Emotion Challenge, organised by (part of) the authors, including the description of the Challenge's database, Sub-Challenges, participants and their approaches, the winners, and the fusion of results to the actual learnt lessons before we finally address the ever-lasting problems and future promising attempts.