Audio genre classification using percussive pattern clustering combined with timbral features

  • Authors:
  • Emiru Tsunoo;George Tzanetakis;Nobutaka Ono;Shigeki Sagayama

  • Affiliations:
  • Graduate School of Information Science and Technology, The University of Tokyo, Japan;Computer Science Department, University of Victoria, Canada;Graduate School of Information Science and Technology, The University of Tokyo, Japan;Graduate School of Information Science and Technology, The University of Tokyo, Japan

  • Venue:
  • ICME'09 Proceedings of the 2009 IEEE international conference on Multimedia and Expo
  • Year:
  • 2009

Quantified Score

Hi-index 0.01

Visualization

Abstract

Many musical genres and styles are characterized by distinct representative rhythmic patterns. In most automatic genre classification systems global statistical features based on timbral dynamics such as Mel-Frequency Cepstral Coefficients (MFCC) are utilized but so far rhythmic information has not so effectively been used. In order to extract bar-long unit rhythmic patterns for a music collection we propose a clustering method based on one-pass dynamic programming and k-means clustering. After extracting the fundamental rhythmic patterns for each style/genre a pattern occurrence histogram is calculated and used as a feature vector for supervised learning. Experimental results show that the automatically calculated rhythmic pattern information can be used to effectively classify musical genre/style and improve upon current approaches based on timbral features.