Incrementally biasing visual search using natural language input

  • Authors:
  • Evan Krause;Rehj Cantrell;Ekaterina Potapova;Michael Zillich;Matthias Scheutz

  • Affiliations:
  • Tufts University, Medford, MA, USA;Indiana University, Bloomington, IN, USA;Vienna University of Technology, Vienna, Austria;Vienna University of Technology, Vienna, Austria;Tufts University, Medford, MA, USA

  • Venue:
  • Proceedings of the 2013 international conference on Autonomous agents and multi-agent systems
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Humans expect interlocutors both human and robot to resolve spoken references to visually-perceivable objects incrementally as the referents are verbally described. For this reason, tight integration of visual search with natural language processing, and real-time operation of both are requirements for natural interactions between humans and robots. In this paper, we present an integrated robotic architecture with novel incremental vision and natural language processing. We demonstrate that incrementally refining attentional focus using linguistic constraints achieves significantly better performance of the vision system compared to non-incremental visual processing.