Neural networks for pattern recognition
Neural networks for pattern recognition
Object Recognition with Features Inspired by Visual Cortex
CVPR '05 Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Volume 2 - Volume 02
Hi-index | 0.00 |
The aim of this paper is to introduce a novel, biologically inspired approach to extract visual features relevant for controlling and understanding reachto-grasp actions. One of the most relevant of such features has been found to be the grip-size defined as the index finger-tip - thumb-tip distance. For this reason, in this paper we focus on this feature. The human visual system is naturally able to recognize many hand configurations - e.g. gestures or different types of grasps - without being affected substantially by the (observer) viewpoint. The proposed computational model preserves this nice ability. It is very likely that this ability may play a crucial role in action understanding within primates (and thus human beings). More specifically, a family of neurons in macaque's ventral premotor area F5 have been discovered which are highly active in correlation with a series of grasp-like movements. This findings triggered a fierce debate about imitation and learning, and inspired several computational models among which the most detailed is due to Oztop and Arbib (MNS model). As a variant of the MNS model, in a previous paper, we proposed the MEP model which relies on an expected perception mechanism. However, both models assume the existence of a mechanism to extract visual features in a viewpoint independent way but neither of them faces the problem of how this mechanism can be achieved in a biologically plausible way. In this paper we propose a neural network model for the extraction of visual features in a viewpoint independent manner, which is based on the work by Poggio and Riesenhuber.