LISTEN: A System for Locating and Tracking Individual Speakers

  • Authors:
  • M. Collobert;R. Feraud;G. Le Tourneur;O. Bernier;J. E. Viallet;Y. Mahieux;D. Collobert

  • Affiliations:
  • -;-;-;-;-;-;-

  • Venue:
  • FG '96 Proceedings of the 2nd International Conference on Automatic Face and Gesture Recognition (FG '96)
  • Year:
  • 1996

Quantified Score

Hi-index 0.00

Visualization

Abstract

Both visual and acoustical informations provide effective means of telecommunication between persons. In this context, the face is the most important part of the person both visually and acoustically. We describe how the cooperation of image and audio processing allows to track a person's face and to collect the audio information it produces. We present detection techniques of regions of interest (e.g. moving regions of skin color), coupled with a neural network based face detector with a low false alarm rate, to locate and track faces. The system is connected to a nine microphone array adaptive beamforming which performs immediate beamforming. Visual and acoustical informations from the speaker face are thus obtained in real time.