Context aware addressee estimation for human robot interaction

  • Authors:
  • Samira Sheikhi;Dinesh Babu Jayagopi;Vasil Khalidov;Jean-Marc Odobez

  • Affiliations:
  • Idiap Research Institute, Martigny, Switzerland;Idiap Research Institute, Martigny, Switzerland;Idiap Research Institute, Martigny, Switzerland;Idiap Research Institute, Martigny, Switzerland

  • Venue:
  • Proceedings of the 6th workshop on Eye gaze in intelligent human machine interaction: gaze in multimodal interaction
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

The paper investigates the problem of addressee recognition -to whom a speaker's utterance is intended- in a setting involving a humanoid robot interacting with multiple persons. More specifically, as it is well known that addressee can primarily be derived from the speaker's visual focus of attention (VFOA) defined as whom or what a person is looking at, we address the following questions: how much does the performance degrade when using automatically extracted VFOA from head pose instead of the VFOA ground-truth? Can the conversational context improve addressee recognition by using it either directly as a side cue in the addressee classifier, or indirectly by improving the VFOA recognition, or in both ways? Finally, from a computational perspective, which VFOA features and normalizations are better and does it matter whether the VFOA recognition module only monitors whether a person looks at potential addressee targets (the robot, people) or if it also considers objects of interest in the environment (paintings in our case) as additional VFOA targets? Experiments on the public Vernissage database where the humanoid Nao robots make a quiz to two participants shows that reducing VFOA confusion (either through context, or by ignoring VFOA targets) improves addressee recognition.