Analysis of multimodal sequences using geometric video representations

  • Authors:
  • Gianluca Monaci;Òscar Divorra Escoda;Pierre Vandergheynst

  • Affiliations:
  • École Polytechnique Fédérale de Lausanne(EPFL), Signal Processing Institute, Lausanne, Switzerland;École Polytechnique Fédérale de Lausanne(EPFL), Signal Processing Institute, Lausanne, Switzerland;École Polytechnique Fédérale de Lausanne(EPFL), Signal Processing Institute, Lausanne, Switzerland

  • Venue:
  • Signal Processing - Special section: Multimodal human-computer interfaces
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents a novel method to correlate audio and visual data generated by the same physical phenomenon, based on sparse geometric representation of video sequences. The video signal is modeled as a sum of geometric primitives evolving through time, that jointly describe the geometric and motion content of the scene. The displacement through time of relevant visual features, like the mouth of a speaker, can thus be compared with the evolution of an audio feature to assess the correspondence between acoustic and visual signals. Experiments show that the proposed approach allows to localize and track the speaker's mouth when several persons are present on the scene, in presence of distracting motion, and without prior face or mouth detection.