Audio-based emotion recognition from natural conversations based on co-occurrence matrix and frequency domain energy distribution features

  • Authors:
  • Aya Sayedelahl;Pouria Fewzee;Mohamed S. Kamel;Fakhri Karray

  • Affiliations:
  • Pattern Analysis and Machine Intelligence Lab, Electrical and Computer Engineering, University of Waterloo, Canada;Pattern Analysis and Machine Intelligence Lab, Electrical and Computer Engineering, University of Waterloo, Canada;Pattern Analysis and Machine Intelligence Lab, Electrical and Computer Engineering, University of Waterloo, Canada;Pattern Analysis and Machine Intelligence Lab, Electrical and Computer Engineering, University of Waterloo, Canada

  • Venue:
  • ACII'11 Proceedings of the 4th international conference on Affective computing and intelligent interaction - Volume Part II
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Emotion recognition from natural speech is a very challenging problem. The audio sub-challenge represents an initial step towards building an efficient audio-visual based emotion recognition system that can detect emotions for real life applications (i.e. human-machine interaction and/or communication). The SEMAINE database, which consists of emotionally colored conversations, is used as the benchmark database. This paper presents our emotion recognition system from speech information in terms of positive/negative valence, and high and low arousal, expectancy and power. We introduce a new set of features including Co-Occurrence matrix based features as well as frequency domain energy distribution based features. Comparisons between well-known prosodic and spectral features and the new features are presented. Classification using the proposed features has shown promising results compared to the classical features on both the development and test data sets.