Audiovisual integration with Segment Models for tennis video parsing

  • Authors:
  • Manolis Delakis;Guillaume Gravier;Patrick Gros

  • Affiliations:
  • IRISA/University of Rennes 1, Campus de Beaulieu, 35042 Rennes, France;IRISA/CNRS, Campus de Beaulieu, 35042 Rennes, France;IRISA/INRIA, Campus de Beaulieu, 35042 Rennes, France

  • Venue:
  • Computer Vision and Image Understanding
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

Automatic video content analysis is an emerging research subject with numerous applications to large video databases and personal video recording systems. The aim of this study is to fuse multimodal information in order to automatically parse the underlying structure of tennis broadcasts. The frame-based observation distributions of Hidden Markov Models are too strict in modeling heterogeneous audiovisual data. We propose instead the use of segmental features, of the framework of Segment Models, to overcome this limitation and extend the synchronization points to the segment boundaries. Considering each segment as a video scene, auditory and visual features collected inside the scene boundaries can thus be sampled and modeled with their native sampling rates and models. Experimental results on a corpus of 15-h tennis video demonstrated a performance superiority of Segment Models with synchronous audiovisual fusion over Hidden Markov Models. Results though with asynchronous fusion are less optimistic.