A real-time, multimodal, and dimensional affect recognition system

  • Authors:
  • Nicole Nielsen Lee;Jocelynn Cu;Merlin Teodosia Suarez

  • Affiliations:
  • Center for Empathic Human-Computer Interactions, De La Salle University, Philippines;Center for Empathic Human-Computer Interactions, De La Salle University, Philippines;Center for Empathic Human-Computer Interactions, De La Salle University, Philippines

  • Venue:
  • PRICAI'12 Proceedings of the 12th Pacific Rim international conference on Trends in Artificial Intelligence
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

This study focuses on the development of a real-time automatic affect recognition system. It adapts a multimodal approach, where affect information taken from two modalities are combined to arrive at an emotion label that is represented in a valence-arousal space. The SEMAINE Database was used to build the affect model. Prosodic and spectral features were used to predict affect from the voice. Temporal templates called Motion History Images (MHI) were used to predict affect from the face. Prediction results from the face and voice models were combined using decision-level fusion. Using support vector machine for regression (SVR), the system was able to correctly identify affect label with a root mean square error (RMSE) of 0.2899 for arousal, and 0.2889 for valence.