A Model of Saliency-Based Visual Attention for Rapid Scene Analysis
IEEE Transactions on Pattern Analysis and Machine Intelligence
Neural Networks: A Comprehensive Foundation
Neural Networks: A Comprehensive Foundation
Contour and Texture Analysis for Image Segmentation
International Journal of Computer Vision
Classifying Images of Materials: Achieving Viewpoint and Illumination Independence
ECCV '02 Proceedings of the 7th European Conference on Computer Vision-Part III
Stochastic Guided Search Model for Search Asymmetries in Visual Search Tasks
BMCV '02 Proceedings of the Second International Workshop on Biologically Motivated Computer Vision
Models of bottom-up and top-down visual attention
Models of bottom-up and top-down visual attention
Object Detection using Background Context
ICPR '04 Proceedings of the Pattern Recognition, 17th International Conference on (ICPR'04) Volume 3 - Volume 03
VOCUS: A Visual Attention System for Object Detection and Goal-Directed Search (Lecture Notes in Computer Science / Lecture Notes in Artificial Intelligence)
The eigen-transform and applications
ACCV'06 Proceedings of the 7th Asian conference on Computer Vision - Volume Part I
Salient region detection using weighted feature maps based on the human visual attention model
PCM'04 Proceedings of the 5th Pacific Rim Conference on Advances in Multimedia Information Processing - Volume Part II
Integrating Visual Context and Object Detection within a Probabilistic Framework
Attention in Cognitive Systems
Hi-index | 0.00 |
Attention plays an important role in human processing of sensory information as a mean of focusing resources toward the most important inputs at the moment. It has in particular been shown to be a key component of vision. In vision it has been argued that the attentional processes are crucial for dealing with the complexity of real world scenes. The problem has often been posed in terms of visual search tasks. It has been shown that both the use of prior task and context information - top-down influences - and favoring information that stands out clearly in the visual field - bottom-up influences - can make such search more efficient. In a generic scene analysis situation one presumably has a combination of these influences and a computational model for visual attention should therefore contain a mechanism for their integration. Such models are abundant for human vision, but relatively few attempts have been made to define any that apply to computer vision.In this article we describe a model that performs such a combination in a principled way. The system learns an optimal representation of the influences of task and context and thereby constructs a biased saliency map representing the top-down information. This map is combined with bottom-up saliency maps in a process evolving over time as a function over the input. The system is applied to search tasks in single images as well as in real scenes, in the latter case using an active vision system capable of shifting its gaze. The proposed model is shown to have desired qualities and to go beyond earlier proposed systems.