Evaluation of facial direction estimation from cameras for multi-modal spoken dialog system

  • Authors:
  • Akihiro Kobayashi;Kentaro Kayama;Etsuo Mizukami;Teruhisa Misu;Hideki Kashioka;Hisashi Kawai;Satoshi Nakamura

  • Affiliations:
  • National Institute of Communications and Technology, Japan;National Institute of Communications and Technology, Japan;National Institute of Communications and Technology, Japan;National Institute of Communications and Technology, Japan;National Institute of Communications and Technology, Japan;National Institute of Communications and Technology, Japan;National Institute of Communications and Technology, Japan

  • Venue:
  • IWSDS'10 Proceedings of the Second international conference on Spoken dialogue systems for ambient environments
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents the results of an evaluation of imageprocessing techniques for estimating facial direction from a camera for a multi-modal spoken dialog system on a large display panel. The system is called the "proactive dialog system" and aims to present acceptable information in an acceptable time. It can detect non-verbal information, such as changes in gaze and facial direction as well as head gestures of the user during dialog, and recommend suitable information. We implemented a dialog scenario to present sightseeing information on the system. Experiments which consist of 100 sesions with 80 subjects were conducted to evaluate the system's efficiency. The system grows particularly clear when dialog contains recommendations.