Human emotion recognition from videos using spatio-temporal and audio features

  • Authors:
  • Munaf Rashid;S. A. Abu-Bakar;Musa Mokji

  • Affiliations:
  • Computer Vision, Video and Image Processing Lab (CVVIP), Faculty of Electrical Engineering, Universiti Teknologi Malaysia, Johor Bahru, Malaysia UTM 81310 and College of Engineering (COE), Karachi ...;Computer Vision, Video and Image Processing Lab (CVVIP), Faculty of Electrical Engineering, Universiti Teknologi Malaysia, Johor Bahru, Malaysia UTM 81310;Computer Vision, Video and Image Processing Lab (CVVIP), Faculty of Electrical Engineering, Universiti Teknologi Malaysia, Johor Bahru, Malaysia UTM 81310

  • Venue:
  • The Visual Computer: International Journal of Computer Graphics
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we present human emotion recognition systems based on audio and spatio-temporal visual features. The proposed system has been tested on audio visual emotion data set with different subjects for both genders. The mel-frequency cepstral coefficient (MFCC) and prosodic features are first identified and then extracted from emotional speech. For facial expressions spatio-temporal features are extracted from visual streams. Principal component analysis (PCA) is applied for dimensionality reduction of the visual features and capturing 97 % of variances. Codebook is constructed for both audio and visual features using Euclidean space. Then occurrences of the histograms are employed as input to the state-of-the-art SVM classifier to realize the judgment of each classifier. Moreover, the judgments from each classifier are combined using Bayes sum rule (BSR) as a final decision step. The proposed system is tested on public data set to recognize the human emotions. Experimental results and simulations proved that using visual features only yields on average 74.15 % accuracy, while using audio features only gives recognition average accuracy of 67.39 %. Whereas by combining both audio and visual features, the overall system accuracy has been significantly improved up to 80.27 %.