Annotation-based multimedia summarization and translation

  • Authors:
  • Katashi Nagao;Shigeki Ohira;Mitsuhiro Yoneoka

  • Affiliations:
  • Nagoya University and CREST, JST, Furo-cho, Chikusa-ku, Nagoya, Japan;Waseda University, Shinjuku-ku, Tokyo, Japan;Tokyo Institute of Technology, Meguro-ku, Tokyo, Japan

  • Venue:
  • COLING '02 Proceedings of the 19th international conference on Computational linguistics - Volume 1
  • Year:
  • 2002

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents techniques for multimedia annotation and their application to video summarization and translation. Our tool for annotation allows users to easily create annotation including voice transcripts, video scene descriptions, and visual/auditory object descriptions. The module for voice transcription is capable of multilingual spoken language identification and recognition. A video scene description consists of semi-automatically detected keyframes of each scene in a video clip and time codes of scenes. A visual object description is created by tracking and interactive naming of people and objects in video scenes. The text data in the multimedia annotation are syntactically and semantically structured using linguistic annotation. The proposed multimedia summarization works upon a multimodal document that consists of a video, keyframes of scenes, and transcripts of the scenes. The multimedia translation automatically generates several versions of multimedia content in different languages.