Multi-stream confidence analysis for audio-visual affect recognition

  • Authors:
  • Zhihong Zeng;Jilin Tu;Ming Liu;Thomas S. Huang

  • Affiliations:
  • Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, IL;Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, IL;Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, IL;Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, IL

  • Venue:
  • ACII'05 Proceedings of the First international conference on Affective Computing and Intelligent Interaction
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

Changes in a speaker’s emotion are a fundamental component in human communication. Some emotions motivate human actions while others add deeper meaning and richness to human interactions. In this paper, we explore the development of a computing algorithm that uses audio and visual sensors to recognize a speaker’s affective state. Within the framework of Multi-stream Hidden Markov Model (MHMM), we analyze audio and visual observations to detect 11 cognitive/emotive states. We investigate the use of individual modality confidence measures as a means of estimating weights when combining likelihoods in the audio-visual decision fusion. Person-independent experimental results from 20 subjects in 660 sequences suggest that the use of stream exponents estimated on training data results in classification accuracy improvement of audio-visual affect recognition.