Grasping of Static and Moving Objects Using a Vision-Based Control Approach

  • Authors:
  • Christopher E. Smith;Nikolaos P. Papanikolopoulos

  • Affiliations:
  • Department of Computer Science and Engineering, University of Colorado at Denver, Campus Box 109, P.O. Box 173364, Denver, CO, U.S.A.;Department of Computer Science, University of Minnesota, 200 Union St. SE, Minneapolis, MN 55455, U.S.A.

  • Venue:
  • Journal of Intelligent and Robotic Systems
  • Year:
  • 1997

Quantified Score

Hi-index 0.00

Visualization

Abstract

Robotic systems require the use of sensing to enable flexible operation in uncalibrated or partially calibrated environments. Recent work combining robotics with vision has emphasized an active vision paradigm where the system changes the pose of the camera to improve environmental knowledge or to establish and preserve a desired relationship between the robot and objects in the environment. Much of this work has concentrated upon the active observation of objects by the robotic agent. We address the problem of robotic visual grasping (eye-in-hand configuration) of static and moving rigid targets. The objective is to move the image projections of certain feature points of the target to effect a vision-guided reach and grasp. An adaptive control algorithm for repositioning a camera compensates for the servoing errors and the computational delays that are introduced by the vision algorithms. Stability issues along with issues concerning the minimum number of required feature points are discussed. Experimental results are presented to verify the validity and the efficacy of the proposed control algorithms. We then address an adaptation to the control paradigm that focuses upon the autonomous grasping of a static or moving object in the manipulator’s workspace. Our work extends the capabilities of an eye-in-hand system beyond those as a ‘pointer’ or a ‘camera orienter’ to provide the flexibility required to robustly interact with the environment in the presence of uncertainty. The proposed work is experimentally verified using the Minnesota Robotic Visual Tracker (MRVT) [7] to automatically select object features, to derive estimates of unknown environmental parameters, and to supply a control vector based upon these estimates to guide the manipulator in the grasping of a static or moving object.