On the limited memory BFGS method for large scale optimization
Mathematical Programming: Series A and B
Unsupervised Learning of Finite Mixture Models
IEEE Transactions on Pattern Analysis and Machine Intelligence
Object Recognition as Machine Translation: Learning a Lexicon for a Fixed Image Vocabulary
ECCV '02 Proceedings of the 7th European Conference on Computer Vision-Part IV
Automatic image annotation and retrieval using cross-media relevance models
Proceedings of the 26th annual international ACM SIGIR conference on Research and development in informaion retrieval
Proceedings of the 26th annual international ACM SIGIR conference on Research and development in informaion retrieval
Experimental result analysis for a generative probabilistic image retrieval model
Proceedings of the 26th annual international ACM SIGIR conference on Research and development in informaion retrieval
Formulating Semantic Image Annotation as a Supervised Learning Problem
CVPR '05 Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Volume 2 - Volume 02
Multiple Bernoulli relevance models for image and video annotation
CVPR'04 Proceedings of the 2004 IEEE computer society conference on Computer vision and pattern recognition
Automated image annotation using global features and robust nonparametric density estimation
CIVR'05 Proceedings of the 4th international conference on Image and Video Retrieval
Image classification for content-based indexing
IEEE Transactions on Image Processing
Information-theoretic semantic multimedia indexing
Proceedings of the 6th ACM international conference on Image and video retrieval
A discrete direct retrieval model for image and video retrieval
CIVR '08 Proceedings of the 2008 international conference on Content-based image and video retrieval
Semantics extraction from images
Knowledge-driven multimedia information extraction and ontology evolution
Hi-index | 0.00 |
This paper is about automatically annotating images with keywords in order to be able to retrieve images with text searches. Our approach is to model keywords such as 'mountain' and 'city' in terms of visual features that were extracted from images. In contrast to other algorithms, each specific keyword-model considers not only its own training data but also the whole training set by utilizing correlations of visual features to refine its own model. Initially, the algorithm clusters all visual features extracted from the full imageset, captures its salient structure (e.g. mixture of clusters or patterns) and represents this as a generic codebook. Then keywords that were associated with images in the training set are encoded as a linear combination of patterns from the generic codebook. We evaluate the validity of our approach in an image retrieval scenario with two distinct large datasets of real-world photos and corresponding manual annotations.