Multimodal speaker verification using ear image features extracted by PCA and ICA

  • Authors:
  • Koji Iwano;Taro Miyazaki;Sadaoki Furui

  • Affiliations:
  • Department of Computer Science, Tokyo Institute of Technology, Tokyo, Japan;Department of Computer Science, Tokyo Institute of Technology, Tokyo, Japan;Department of Computer Science, Tokyo Institute of Technology, Tokyo, Japan

  • Venue:
  • AVBPA'05 Proceedings of the 5th international conference on Audio- and Video-Based Biometric Person Authentication
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper first compares performances of two authentication methods using ear images, in which feature vectors are extracted by either principal component analysis (PCA) or independent component analysis (ICA). Next, the effectiveness of combining PCA- and ICA-based ear authentication methods is investigated. In our previous work, we proposed an audio-visual person authentication using speech and ear images with the aim of increasing noise robustness in mobile environments. In this paper, we apply the best ear authentication method to our audio-visual authentication method and examine its robustness. Experiments were conducted using an audio-visual database collected from 36 male speakers in five sessions over a half year. Speech data were contaminated with white noise at various SNR conditions. Experimental results show that: (1) PCA outperforms ICA in the ear authentication framework using GMMs; (2) the fusion of PCA- and ICA-based ear authentication is effective; and (3) by combining the fusion method for ear images with the speech-based method, person authentication performance can be improved. The audio-visual person authentication method achieves better performance than ear-based as well as speech-based methods in an SNR range between 15 and 30dB.