Optimized discriminative transformations for speech features based on minimum classification error

  • Authors:
  • Behzad Zamani;Ahmad Akbari;Babak Nasersharif;Azarakhsh Jalalvand

  • Affiliations:
  • Audio & Speech Processing Lab, Computer Engineering Department, Iran University of Science & Technology, Iran and Information Technology Research Center, Iran University of Science & Technology, I ...;Audio & Speech Processing Lab, Computer Engineering Department, Iran University of Science & Technology, Iran and Information Technology Research Center, Iran University of Science & Technology, I ...;Audio & Speech Processing Lab, Computer Engineering Department, Iran University of Science & Technology, Iran and Computer Engineering Department, Faculty of Engineering, University of Guilan, Ira ...;Audio & Speech Processing Lab, Computer Engineering Department, Iran University of Science & Technology, Iran and Information Technology Research Center, Iran University of Science & Technology, I ...

  • Venue:
  • Pattern Recognition Letters
  • Year:
  • 2011

Quantified Score

Hi-index 0.10

Visualization

Abstract

Feature extraction is an important component of pattern classification and speech recognition. Extracted features should discriminate classes from each other while being robust to environmental conditions such as noise. For this purpose, several feature transformations are proposed which can be divided into two main categories: data-dependent transformation and classifier-dependent transformation. The drawback of data-dependent transformation is that its optimization criteria are different from the measure of classification error which can potentially degrade the classifier's performance. In this paper, we propose a framework to optimize data-dependent feature transformations such as PCA (Principal Component Analysis), LDA (Linear Discriminant Analysis) and HLDA (Heteroscedastic LDA) using minimum classification error (MCE) as the main objective. The classifier itself is based on Hidden Markov Model (HMM). In our proposed HMM minimum classification error technique, the transformation matrices are modified to minimize the classification error for the mapped features, and the dimension of the feature vector is not changed. To evaluate the proposed methods, we conducted several experiments on the TIMIT phone recognition and the Aurora2 isolated word recognition tasks. The experimental results show that the proposed methods improve performance of PCA, LDA and HLDA transformation for mapping Mel-frequency cepstral coefficients (MFCC).