A neural network model for a view independent extraction of reach-to-grasp action features

  • Authors:
  • Roberto Prevete;Matteo Santoro;Ezio Catanzariti;Giovanni Tessitore

  • Affiliations:
  • Department of Physical Sciences, University of Naples Federico II, Naples, Italy;Department of Physical Sciences, University of Naples Federico II, Naples, Italy;Department of Physical Sciences, University of Naples Federico II, Naples, Italy;Department of Physical Sciences, University of Naples Federico II, Naples, Italy

  • Venue:
  • BVAI'07 Proceedings of the 2nd international conference on Advances in brain, vision and artificial intelligence
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

The aim of this paper is to introduce a novel, biologically inspired approach to extract visual features relevant for controlling and understanding reachto-grasp actions. One of the most relevant of such features has been found to be the grip-size defined as the index finger-tip - thumb-tip distance. For this reason, in this paper we focus on this feature. The human visual system is naturally able to recognize many hand configurations - e.g. gestures or different types of grasps - without being affected substantially by the (observer) viewpoint. The proposed computational model preserves this nice ability. It is very likely that this ability may play a crucial role in action understanding within primates (and thus human beings). More specifically, a family of neurons in macaque's ventral premotor area F5 have been discovered which are highly active in correlation with a series of grasp-like movements. This findings triggered a fierce debate about imitation and learning, and inspired several computational models among which the most detailed is due to Oztop and Arbib (MNS model). As a variant of the MNS model, in a previous paper, we proposed the MEP model which relies on an expected perception mechanism. However, both models assume the existence of a mechanism to extract visual features in a viewpoint independent way but neither of them faces the problem of how this mechanism can be achieved in a biologically plausible way. In this paper we propose a neural network model for the extraction of visual features in a viewpoint independent manner, which is based on the work by Poggio and Riesenhuber.