The multimodal presentation dashboard

  • Authors:
  • Michael Johnston;Patrick Ehlen;David Gibbon;Zhu Liu

  • Affiliations:
  • AT&T Labs Research, Florham Park, NJ;Stanford University, Palo Alto, CA;AT&T Labs Research, Florham Park, NJ;AT&T Labs Research, Florham Park, NJ

  • Venue:
  • NAACL-HLT-Dialog '07 Proceedings of the Workshop on Bridging the Gap: Academic and Industrial Research in Dialog Technologies
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

The multimodal presentation dashboard allows users to control and browse presentation content such as slides and diagrams through a multimodal interface that supports speech and pen input. In addition to control commands (e.g. "take me to slide 10"), the system allows multimodal search over content collections. For example, if the user says "get me a slide about internet telephony," the system will present a ranked series of candidate slides that they can then select among using voice, pen, or a wireless remote. As presentations are loaded, their content is analyzed and language and understanding models are built dynamically. This approach frees the user from the constraints of linear order allowing for a more dynamic and responsive presentation style.