Working with robots and objects: revisiting deictic reference for achieving spatial common ground

  • Authors:
  • Andrew G. Brooks;Cynthia Breazeal

  • Affiliations:
  • MIT Media Laboratory, Cambridge, MA;MIT Media Laboratory, Cambridge, MA

  • Venue:
  • Proceedings of the 1st ACM SIGCHI/SIGART conference on Human-robot interaction
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

Robust joint visual attention is necessary for achieving a common frame of reference between humans and robots interacting multimodally in order to work together on real-world spatial tasks involving objects. We make a comprehensive examination of one component of this process that is often otherwise implemented in an ad hoc fashion: the ability to correctly determine the object referent from deictic reference including pointing gestures and speech. From this we describe the development of a modular spatial reasoning framework based around decomposition and resynthesis of speech and gesture into a language of pointing and object labeling. This framework supports multimodal and unimodal access in both real-world and mixed-reality workspaces, accounts for the need to discriminate and sequence identical and proximate objects, assists in overcoming inherent precision limitations in deictic gesture, and assists in the extraction of those gestures. We further discuss an implementation of the framework that has been deployed on two humanoid robot platforms to date.