Object recognition and full pose registration from a single image for robotic manipulation

  • Authors:
  • Alvaro Collet;Dmitry Berenson;Siddhartha S. Srinivasa;Dave Ferguson

  • Affiliations:
  • The Robotics Institute, Carnegie Mellon University, Pittsburgh, PA;The Robotics Institute, Carnegie Mellon University, Pittsburgh, PA;Intel Research Pittsburgh, Pittsburgh, PA;Intel Research Pittsburgh, Pittsburgh, PA

  • Venue:
  • ICRA'09 Proceedings of the 2009 IEEE international conference on Robotics and Automation
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Robust perception is a vital capability for robotic manipulation in unstructured scenes. In this context, full pose estimation of relevant objects in a scene is a critical step towards the introduction of robots into household environments. In this paper, we present an approach for building metric 3D models of objects using local descriptors from several images. Each model is optimized to fit a set of calibrated training images, thus obtaining the best possible alignment between the 3D model and the real object. Given a new test image, we match the local descriptors to our stored models online, using a novel combination of the RANSAC and Mean Shift algorithms to register multiple instances of each object. A robust initialization step allows for arbitrary rotation, translation and scaling of objects in the test images. The resulting system provides markerless 6-DOF pose estimation for complex objects in cluttered scenes. We provide experimental results demonstrating orientation and translation accuracy, as well a physical implementation of the pose output being used by an autonomous robot to perform grasping in highly cluttered scenes.