Object Recognition Using Multidimensional Receptive Field Histograms
ECCV '96 Proceedings of the 4th European Conference on Computer Vision-Volume I - Volume I
A Binocular Stereo Algorithm for Log-Polar Foveated Systems
BMCV '02 Proceedings of the Second International Workshop on Biologically Motivated Computer Vision
Distinctive Image Features from Scale-Invariant Keypoints
International Journal of Computer Vision
Perceptual anchoring of symbols for action
IJCAI'01 Proceedings of the 17th international joint conference on Artificial intelligence - Volume 1
From pixels to objects: enabling a spatial model for humanoid social robots
ICRA'09 Proceedings of the 2009 IEEE international conference on Robotics and Automation
SURF: speeded up robust features
ECCV'06 Proceedings of the 9th European conference on Computer Vision - Volume Part I
Learning Object Affordances: From Sensory--Motor Coordination to Imitation
IEEE Transactions on Robotics
IEEE Transactions on Evolutionary Computation
Hi-index | 0.00 |
This paper presents an approach to endow a humanoid robot with the capability of learning new objects and recognizing them in an unstructured environment. New objects are learnt, whenever an unrecognized one is found within a certain (small) distance from the robot head. Recognized objects are mapped to an ego-centric frame of reference, which together with a simple short-term memory mechanism, makes this mapping persistent. This allows the robot to be aware of their presence even if temporarily out of the field of view, thus providing a primary spatial model of the environment (as far as known objects are concerned). SIFT features are used, not only for recognizing previously learnt objects, but also to allow the robot to estimate their distance (depth perception). The humanoid platform used for the experiments was the iCub humanoid robot. This capability functions together with iCub's low-level attention system: recognized objects enact salience thus attracting the robot attention, by gazing at them, each one in turn. We claim that the presented approach is a contribution towards linking a bottom-up attention system with top-down cognitive information.