Head gestures for perceptual interfaces: The role of context in improving recognition

  • Authors:
  • Louis-Philippe Morency;Candace Sidner;Christopher Lee;Trevor Darrell

  • Affiliations:
  • MIT CSAIL, Cambridge, MA 02139, USA;BAE Systems AIT, Burlington, MA 01803, USA;Boston Dynamics, Waltham, MA 02139, USA;MIT CSAIL, Cambridge, MA 02139, USA

  • Venue:
  • Artificial Intelligence
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

Head pose and gesture offer several conversational grounding cues and are used extensively in face-to-face interaction among people. To accurately recognize visual feedback, humans often use contextual knowledge from previous and current events to anticipate when feedback is most likely to occur. In this paper we describe how contextual information can be used to predict visual feedback and improve recognition of head gestures in human-computer interfaces. Lexical, prosodic, timing, and gesture features can be used to predict a user's visual feedback during conversational dialog with a robotic or virtual agent. In non-conversational interfaces, context features based on user-interface system events can improve detection of head gestures for dialog box confirmation or document browsing. Our user study with prototype gesture-based components indicate quantitative and qualitative benefits of gesture-based confirmation over conventional alternatives. Using a discriminative approach to contextual prediction and multi-modal integration, performance of head gesture detection was improved with context features even when the topic of the test set was significantly different than the training set.