The use of Speech and Lip Modalities for Robust Speaker Verification under Adverse Conditions

  • Authors:
  • T. Wark;S. Sridharan;V. Chandran

  • Affiliations:
  • Queensland University of Technology;Queensland University of Technology;Queensland University of Technology

  • Venue:
  • ICMCS '99 Proceedings of the IEEE International Conference on Multimedia Computing and Systems - Volume 2
  • Year:
  • 1999

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper investigates the use of lip information, in conjunction with speech information, for robust speaker verification in the presence of background noise. It has been previously shown in our own work, and in the work of others, that features extracted from a speaker's moving lips hold speaker dependencies which are complementary with speech features. We demonstrate that the fusion of lip and speech information allows for a highly robust speaker verification system which outperforms the performance of either sub-system. We present a new technique for determining the weighting to be applied to each modality so as to optimize the performance of the fused system. Given a correct weighting, lip information is shown to be highly effective for reducing the false acceptance and false rejection error rates in the presence of background noise.