Grasping of static and moving objects using a vision-based control approach

  • Authors:
  • Affiliations:
  • Venue:
  • IROS '95 Proceedings of the International Conference on Intelligent Robots and Systems-Volume 1 - Volume 1
  • Year:
  • 1995

Quantified Score

Hi-index 0.00

Visualization

Abstract

Recent work combining robotics with vision has emphasized an active vision paradigm where the system changes the pose of the camera to improve environmental knowledge or to establish and presence a desired relationship between the robot and objects in the environment. Much of this work has concentrated upon the active observation of objects by the robotic agent. In this paper we present extensions to the controlled active vision framework that focus upon the autonomous grasping of a moving or static object in the manipulator's workspace. Our work extends the capabilities of an eye-in-hand system beyond those as a "pointer" or a "camera orienter" to provide the flexibility required to robustly interact with the environment in the presence of uncertainty. The proposed work is experimentally verified using the Minnesota robotic visual tracker to automatically select object features, to derive estimates of unknown environmental parameters, and to supply a control vector based upon these estimates to guide the manipulator in the grasping of a moving or static object.