Multi-party human-machine interaction using a smart multimodal digital signage

  • Authors:
  • Tony Tung;Randy Gomez;Tatsuya Kawahara;Takashi Matsuyama

  • Affiliations:
  • Academic Center for Computing and Media Studies and Graduate School of Informatics, Kyoto University, Japan;Academic Center for Computing and Media Studies and Graduate School of Informatics, Kyoto University, Japan;Academic Center for Computing and Media Studies and Graduate School of Informatics, Kyoto University, Japan;Academic Center for Computing and Media Studies and Graduate School of Informatics, Kyoto University, Japan

  • Venue:
  • HCI'13 Proceedings of the 15th international conference on Human-Computer Interaction: interaction modalities and techniques - Volume Part IV
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we present a novel multimodal system designed for smooth multi-party human-machine interaction. HCI for multiple users is challenging because simultaneous actions and reactions have to be consistent. Here, the proposed system consists of a digital signage or large display equipped with multiple sensing devices: a 19-channel microphone array, 6 HD video cameras (3 are placed on top and 3 on the bottom of the display), and two depth sensors. The display can show various contents, similar to a poster presentation, or multiple windows (e.g., web browsers, photos, etc.). On the other hand, multiple users positioned in front of the panel can freely interact using voice or gesture while looking at the displayed contents, without wearing any particular device (such as motion capture sensors or head mounted devices). Acoustic and visual information processing are performed jointly using state-of-the-art techniques to obtain individual speech and gaze direction. Hence displayed contents can be adapted to users' interests.