Affective music recommendation system using input images

  • Authors:
  • Shoto Sasaki;Tatsunori Hirai;Hayato Ohya;Shigeo Morishima

  • Affiliations:
  • Waseda University;Waseda University;Waseda University;Waseda University

  • Venue:
  • ACM SIGGRAPH 2013 Posters
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Music that matches our current mood can create a deep impression, which we usually want to enjoy when we listen to music. However, we do not know which music best matches our present mood. We have to listen to each song, searching for music that matches our mood. As it is difficult to select music manually, we need a recommendation system that can operate affectively. Most recommendation methods, such as collaborative filtering or content similarity, do not target a specific mood. In addition, there may be no word exactly specifying the mood. Therefore, textual retrieval is not effective. In this paper, we assume that there exists a relationship between our mood and images because visual information affects our mood when we listen to music. We now present an affective music recommendation system using an input image without textual information.