Augmented Transition Networks as Video Browsing Models for Multimedia Databases and Multimedia Information Systems

  • Authors:
  • Shu-Ching Chen;Srinivas Sista;Mei-Ling Shyu;R. L. Kashyap

  • Affiliations:
  • -;-;-;-

  • Venue:
  • ICTAI '99 Proceedings of the 11th IEEE International Conference on Tools with Artificial Intelligence
  • Year:
  • 1999

Quantified Score

Hi-index 0.00

Visualization

Abstract

In an interactive multimedia information system, users should have the flexibility to browse and choose various scenarios they want to see. This means that two-way communications should be captured by the conceptual model. Digital video has gained increasing popularity in many multimedia applications. Instead of sequential access to the video contents, the structuring and modeling of video data so that users can quickly and easily browse and retrieve interesting materials becomes an important issue in designing multimedia information systems. An abstract semantic model called the augmented transition network (ATN), which can model video data and user interactions, is proposed in this paper. An ATN and its sub-networks can model video data based on different granularities such as scenes, shots and key frames. Multimedia input strings are used as inputs for ATNs. The details of how to use multimedia input strings to model video data are also discussed. Key frame selection is based on temporal and spatial relations of semantic objects in each shot. The temporal and spatial relations of semantic objects are captured from our proposed unsupervised video segmentation method which considers the problem of partitioning each frame as a joint estimation of the partition and class parameter variables. Unlike existing semantic models which only model multimedia presentation, multimedia database searching, or browsing, ATNs together with multimedia input strings can model these three in one framework.