Affective content analysis by mid-level representation in multiple modalities

  • Authors:
  • Min Xu;Suhuai Luo;Jesse S. Jin;Mira Park

  • Affiliations:
  • University of Newcastle, Callaghan, Australia;University of Newcastle, Callaghan, Australia;University of Newcastle, Callaghan, Australia;University of Newcastle, Callaghan, Australia

  • Venue:
  • Proceedings of the First International Conference on Internet Multimedia Computing and Service
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Movie affective content detection attracts ever-increasing research efforts. However, the affective content analysis is still a challenging task due to the gap between low-level perceptual features and high-level human perception of the media. Moreover, clues from multiple modalities should be considered for affective analysis, since they were used in movies to represent emotions and romance emotional atmosphere. In this paper, mid-level representations are generated from low-level features. These mid-level representations are from multiple modalities and used for affective content inference. Besides video shots which is commonly used for video content analysis, audio sounds, dialogue and subtitle are explored to contribute to detect affective content. Since affective analysis rely on movie genres, experiments are implemented in respective genres. The results shows that audio sounds, dialogues and subtitles are effective and efficient for affective content detection.