Using Multi-view Recognition and Meta-data Annotation to Guide a Robot's Attention

  • Authors:
  • Alexander Thomas;Vittorio Ferrari;Bastian Leibe;Tinne Tuytelaars;Luc Van Gool

  • Affiliations:
  • ESAT-PSI/VISICS, KU Leuven, Heverlee 3001, Belgium,;Computer Vision Laboratory, ETH Zürich, Zürich 8092,Switzerland;UMIC Research Centre RWTH Aachen, 52056 Aachen Germany;ESAT-PSI/VISICS, KU Leuven, Heverlee 3001, Belgium;Computer Vision Laboratory, ETH Zürich, Zürich 8092,Switzerland, ESAT-PSI/VISICS, KU Leuven, Heverlee 3001, Belgium

  • Venue:
  • International Journal of Robotics Research
  • Year:
  • 2009

Quantified Score

Hi-index 0.01

Visualization

Abstract

In the transition from industrial to service robotics, robots will have to deal with increasingly unpredictable and variable environments. We present a system that is able to recognize objects of a certain class in an image and to identify their parts for potential interactions. The method can recognize objects from arbitrary viewpoints and generalizes to instances that have never been observed during training, even if they are partially occluded and appear against cluttered backgrounds. Our approach builds on the implicit shape model of Leibe et al. We extend it to couple recognition to the provision of meta-data useful for a task and to the case of multiple viewpoints by integrating it with the dense multi-view correspondence finder of Ferrari et al. Meta-data can be part labels but also depth estimates, information on material types, or any other pixelwise annotation. We present experimental results on wheelchairs, cars, and motorbikes.