Semi-coupled hidden Markov model with state-based alignment strategy for audio-visual emotion recognition

  • Authors:
  • Jen-Chun Lin;Chung-Hsien Wu;Wen-Li Wei

  • Affiliations:
  • Department of Computer Science and Information Engineering, National Cheng Kung University, Taiwan, R.O.C.;Department of Computer Science and Information Engineering, National Cheng Kung University, Taiwan, R.O.C.;Department of Computer Science and Information Engineering, National Cheng Kung University, Taiwan, R.O.C.

  • Venue:
  • ACII'11 Proceedings of the 4th international conference on Affective computing and intelligent interaction - Volume Part I
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents an approach to bi-modal emotion recognition based on a semi-coupled hidden Markov model (SC-HMM). A simplified state-based bi-modal alignment strategy in SC-HMM is proposed to align the temporal relation of states between audio and visual streams. Based on this strategy, the proposed SC-HMM can alleviate the problem of data sparseness and achieve better statistical dependency between states of audio and visual HMMs in most real world scenarios. For performance evaluation, audio-visual signals with four emotional states (happy, neutral, angry and sad) were collected. Each of the invited seven subjects was asked to utter 30 types of sentences twice to generate emotional speech and facial expression for each emotion. Experimental results show the proposed bi-modal approach outperforms other fusion-based bi-modal emotion recognition methods.