Dictionary learning based sparse coefficients for audio classification with max and average pooling

  • Authors:
  • Syed Zubair;Fei Yan;Wenwu Wang

  • Affiliations:
  • Centre for Vision, Speech and Signal Processing, University of Surrey, Guildford, GU2 7XH, UK;Centre for Vision, Speech and Signal Processing, University of Surrey, Guildford, GU2 7XH, UK;Centre for Vision, Speech and Signal Processing, University of Surrey, Guildford, GU2 7XH, UK

  • Venue:
  • Digital Signal Processing
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Audio classification is an important problem in signal processing and pattern recognition with potential applications in audio retrieval, documentation and scene analysis. Common to general signal classification systems, it involves both training and classification (or testing) stages. The performance of an audio classification system, such as its complexity and classification accuracy, depends highly on the choice of the signal features and the classifiers. Several features have been widely exploited in existing methods, such as the mel-frequency cepstrum coefficients (MFCCs), line spectral frequencies (LSF) and short time energy (STM). In this paper, instead of using these well-established features, we explore the potential of sparse features, derived from the dictionary of signal atoms using sparse coding based on e.g. orthogonal matching pursuit (OMP), where the atoms are adapted directly from audio training data using the K-SVD dictionary learning algorithm. To reduce the computational complexity, we propose to perform pooling and sampling operations on the sparse coefficients. Such operations also help to maintain a unified dimension of the signal features, regardless of the various lengths of the training and testing signals. Using the popular support vector machine (SVM) as the classifier, we examine the performance of the proposed classification system for two binary classification problems, namely speech-music classification and male-female speech discrimination and a multi-class problem, speaker identification. The experimental results show that the sparse (max-pooled and average-pooled) coefficients perform better than the classical MFCCs features, in particular, for noisy audio data.