Model-based and empirical evaluation of multimodal interactive error correction

  • Authors:
  • Bernhard Suhm;Brad Myers;Alex Waibel

  • Affiliations:
  • Interactive Systems Laboratories, Carnegie Mellon University/Universität Karlsruhe;Human Computer Interaction Institute, Carnegie Mellon University;Interactive Systems Laboratories, Carnegie Mellon University/Universität Karlsruhe

  • Venue:
  • Proceedings of the SIGCHI conference on Human Factors in Computing Systems
  • Year:
  • 1999

Quantified Score

Hi-index 0.01

Visualization

Abstract

Our research addresses the problem of error correction in speechuser interfaces. Previous work hypothesized that switching modalitycould speed up interactive correction of recognition errors(so-called multimodal error correction). We present a user studythat compares, on a dictation task, multimodal error correctionwith conventional interactive correction, such as speaking again,choosing Tom a list, and keyboard input. Results show thatmultimodal correction is faster than conventional correctionwithout keyboard input, but slower than correction by typing forusers with good typing skills. Furthermore, while users initiallyprefer speech, they learn to avoid ineffective correctionmodalities with experience. To extrapolate results from this userstudy we developed a performance model of multimodal interactionthat predicts input speed including time needed for errorcorrection. We apply the model to estimate the impact ofrecognition technology improvements on correction speeds and theinfluence of recognition accuracy and correction method on theproductivity of dictation systems. Our model is a first steptowards formalizing multimodal (recognition-based) interaction.