Bimodal speaker identification using dynamic bayesian network

  • Authors:
  • Dongdong Li;LiFeng Sang;Yingchun Yang;Zhaohui Wu

  • Affiliations:
  • Department of Computer Science, Zhejiang University, Hang Zhou, P.R China;Department of Computer Science, Zhejiang University, Hang Zhou, P.R China;Department of Computer Science, Zhejiang University, Hang Zhou, P.R China;Department of Computer Science, Zhejiang University, Hang Zhou, P.R China

  • Venue:
  • SINOBIOMETRICS'04 Proceedings of the 5th Chinese conference on Advances in Biometric Person Authentication
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

The authentication of a person requires a consistently high recognition accuracy which is difficult to attain using a single recognition modality This paper assesses the fusion of voiceprint and face feature for bimodal speaker identification using Dynamic Bayesian Network (DBN) Our contribution is to propose a general feature-level fusion framework in bimodal speaker identification Within the framework, the voice and face feature are combined into a single DBN to obtain better performance than any single system alone The tests were conducted on a multi-modal database of 54 users who provided voiceprint and face data of different speech type and content .We compare our approach with mono-modal system and other classic decision-level methods and show that feature-level fusion using dynamic Bayesian network improved performance by about 4-5%, much better than the others.