A Semantic Model for Video Description and Retrieval
PCM '02 Proceedings of the Third IEEE Pacific Rim Conference on Multimedia: Advances in Multimedia Information Processing
Rule-based spatiotemporal query processing for video databases
The VLDB Journal — The International Journal on Very Large Data Bases
Computational approaches to temporal sampling of video sequences
ACM Transactions on Multimedia Computing, Communications, and Applications (TOMCCAP)
Hi-index | 0.00 |
Traditional research on video data retrieval follows two general approaches. One is based on text annotation and the other on content-based comparison. However, these approaches do not fully make use of the meaning implied in a video stream. To improve these approaches, a semantic video model cooperated with a knowledge database is studied. In this paper, we propose a new semantic video model and focus on presenting the semantic meaning implied in a video. According to the granularity of the meaning implied in a video, a five-level layered structure to model a video stream is proposed. A mechanism is also provided to construct the five levels based on the knowledge categories defined in the knowledge database. The five-level layered structure consists of raw-data levels and semantic-data levels. A uniform semantics representation is proposed to represent the semantic-data levels. This uniform semantics representation allows measuring the similarity of two video streams with different duration. Then an interactive interface can provide browsing and querying video data efficiently through the uniform semantics representation.