Multimodal integration-a statistical view

  • Authors:
  • Lizhong Wu;S. L. Oviatt;P. R. Cohen

  • Affiliations:
  • Dept. of Comput. Sci. & Eng., Oregon Graduate Inst. of Sci. & Technol., Portland, OR;-;-

  • Venue:
  • IEEE Transactions on Multimedia
  • Year:
  • 1999

Quantified Score

Hi-index 0.03

Visualization

Abstract

We present a statistical approach to developing multimodal recognition systems and, in particular, to integrating the posterior probabilities of parallel input signals involved in the multimodal system. We first identify the primary factors that influence multimodal recognition performance by evaluating the multimodal recognition probabilities. We then develop two techniques, an estimate approach and a learning approach, which are designed to optimize accurate recognition during the multimodal integration process. We evaluate these methods using Quickset, a speech/gesture multimodal system, and report evaluation results based on an empirical corpus collected with Quickset. From an architectural perspective, the integration technique presented offers enhanced robustness. It also is premised on more realistic assumptions than previous multimodal systems using semantic fusion. From a methodological standpoint, the evaluation techniques that we describe provide a valuable tool for evaluating multimodal systems