Toward Multi-modal Music Emotion Classification

  • Authors:
  • Yi-Hsuan Yang;Yu-Ching Lin;Heng-Tze Cheng;I-Bin Liao;Yeh-Chin Ho;Homer H. Chen

  • Affiliations:
  • National Taiwan University,;National Taiwan University,;National Taiwan University,;Telecommunication Laboratories, Chunghwa Telecom,;Telecommunication Laboratories, Chunghwa Telecom,;National Taiwan University,

  • Venue:
  • PCM '08 Proceedings of the 9th Pacific Rim Conference on Multimedia: Advances in Multimedia Information Processing
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

The performance of categorical music emotion classification that divides emotion into classes and uses audio features alone for emotion classification has reached a limit due to the presence of a semantic gap between the object feature level and the human cognitive level of emotion perception. Motivated by the fact that lyrics carry rich semantic information of a song, we propose a multi-modal approach to help improve categorical music emotion classification. By exploiting both the audio features and the lyrics of a song, the proposed approach improves the 4-class emotion classification accuracy from 46.6% to 57.1%. The results also show that the incorporation of lyrics significantly enhances the classification accuracy of valence.