An efficient unification-based multimodal language processor in multimodal input fusion

  • Authors:
  • Yong Sun;Yu Shi;Fang Chen;Vera Chung

  • Affiliations:
  • National ICT Australia, Eveleigh NSW, Australia and The University of Sydney, Redfern NSW, Australia;National ICT Australia, Eveleigh NSW, Australia;National ICT Australia, Eveleigh NSW, Australia and The University of Sydney, Redfern NSW, Australia;The University of Sydney, Redfern NSW, Australia

  • Venue:
  • OZCHI '07 Proceedings of the 19th Australasian conference on Computer-Human Interaction: Entertaining User Interfaces
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

A Multimodal User Interface (MMUI) allows a user to interact with a computer in a way similar to human-to-human communication, for example, through speech and gesture. Being an essential component in MMUIs, Multimodal Input Fusion should be able to find the semantic interpretation of a user's intention from recognized multimodal symbols which are semantically complementary. We enhanced our efficient unification-based multimodal parsing processor, which has the potential to achieve low polynomial computational complexity while parsing versatile multimodal inputs within a speech and gesture based MMUI, to handle multimodal inputs from more than two modalities. Its ability to disambiguate speech recognition results with gesture recognition results was verified with an experiment. The analysis of experiment results demonstrates that the improvement is significant after applying this technique.