Selecting relevant visual features for speechreading

  • Authors:
  • V. Estellers;M. Gurban;J. P. Thiran

  • Affiliations:
  • Ecole Polytechnique Fédérale de Lausanne, Signal Processing Laboratory 5, Lausanne, Switzerland;Ecole Polytechnique Fédérale de Lausanne, Signal Processing Laboratory 5, Lausanne, Switzerland;Ecole Polytechnique Fédérale de Lausanne, Signal Processing Laboratory 5, Lausanne, Switzerland

  • Venue:
  • ICIP'09 Proceedings of the 16th IEEE international conference on Image processing
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

A quantitative measure of relevance is proposed for the task of constructing visual feature sets which are at the same time relevant and compact. A feature's relevance is given by the amount of information that it contains about the problem, while compactness is achieved by preventing the replication of information between features. To achieve these goals, we use mutual information both for assessing relevance and measuring the redundancy between features. Our application is speechreading, that is, speech recognition performed on the video of the speaker. This is justified by the fact that the performance of audio speech recognition can be improved by augmenting the audio features with visual ones, especially when there is noise in the audio channel. We report significant improvements compared to the most common method of dimensionality reduction for speechreading, Linear Discriminant Analysis (LDA).