Combining multimodal preferences for multimedia information retrieval
Proceedings of the international workshop on Workshop on multimedia information retrieval
A human-centered multiple instance learning framework for semantic video retrieval
IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews
Local feature trajectories for efficient event-based indexing of video sequences
CIVR'06 Proceedings of the 5th international conference on Image and Video Retrieval
Learning user queries in multimodal dissimilarity spaces
AMR'05 Proceedings of the Third international conference on Adaptive Multimedia Retrieval: user, context, and feedback
Hi-index | 0.00 |
This paper deals with the problem of event discrimination in generic video documents. We propose an investigation on the design of an activity-based similarity measure derived from motion analysis. In an unsupervised context, our approach relies on the nonlinear temporal modeling of wavelet-based motion features directly estimated from the video frame. On the basis of the support vector machine (SVM) regression, this nonlinear model is able to learn the behavior of the motion descriptors along the temporal dimension and to capture useful information about the dynamic content of the shot. A similarity measure associated with our temporal model is then defined. This measure defines a metric between video segments according to spatial and temporal properties of the movements and provides a theoretic framework to compare, sort and classify videos. Experiments on a large annotated video database and a comparison with a similarity measure based on motion histograms shows that our approach is effective in discriminating between video events without any prior knowledge.