Comparison of image transform-based features for visual speech recognition in clean and corrupted videos

  • Authors:
  • Rowan Seymour;Darryl Stewart;Ji Ming

  • Affiliations:
  • School of Electronics, Electrical Engineering and Computer Science, Queen's University of Belfast, Belfast, Northern Ireland, UK;School of Electronics, Electrical Engineering and Computer Science, Queen's University of Belfast, Belfast, Northern Ireland, UK;School of Electronics, Electrical Engineering and Computer Science, Queen's University of Belfast, Belfast, Northern Ireland, UK

  • Venue:
  • Journal on Image and Video Processing - Anthropocentric Video Analysis: Tools and Applications
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present results of a study into the performance of a variety of different image transform-based feature types for speaker-independent visual speech recognition of isolated digits. This includes the first reported use of features extracted using a discrete curvelet transform. The study will show a comparison of some methods for selecting features of each feature type and show the relative benefits of both static and dynamic visual features. The performance of the features will be tested on both clean video data and also video data corrupted in a variety of ways to assess each feature type's robustness to potential real-world conditions. One of the test conditions involves a novel form of video corruption we call jitter which simulates camera and/or head movement during recording.