Being bored? Recognising natural interest by extensive audiovisual integration for real-life application

  • Authors:
  • Björn Schuller;Ronald Müller;Florian Eyben;Jürgen Gast;Benedikt Hörnler;Martin Wöllmer;Gerhard Rigoll;Anja Höthker;Hitoshi Konosu

  • Affiliations:
  • Institute for Human-Machine Communication, Technische Universität München, D-80333 München, Germany;Altran Technologies, Bernhard-Wicki-Str. 3, 80636 München, Germany;Institute for Human-Machine Communication, Technische Universität München, D-80333 München, Germany;Institute for Human-Machine Communication, Technische Universität München, D-80333 München, Germany;Institute for Human-Machine Communication, Technische Universität München, D-80333 München, Germany;Institute for Human-Machine Communication, Technische Universität München, D-80333 München, Germany;Institute for Human-Machine Communication, Technische Universität München, D-80333 München, Germany;Toyota Motor Europe, Production Engineering - Advanced Technologies, B-1930 Zaventem, Belgium;Toyota Motor Corporation, 1 Toyota-cho, Toyota City, Aichi 471-8571, Japan

  • Venue:
  • Image and Vision Computing
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Automatic detection of the level of human interest is of high relevance for many technical applications, such as automatic customer care or tutoring systems. However, the recognition of spontaneous interest in natural conversations independently of the subject remains a challenge. Identification of human affective states relying on single modalities only is often impossible, even for humans, since different modalities contain partially disjunctive cues. Multimodal approaches to human affect recognition generally are shown to boost recognition performance, yet are evaluated in restrictive laboratory settings only. Herein we introduce a fully automatic processing combination of Active-Appearance-Model-based facial expression, vision-based eye-activity estimation, acoustic features, linguistic analysis, non-linguistic vocalisations, and temporal context information in an early feature fusion process. We provide detailed subject-independent results for classification and regression of the Level of Interest using Support-Vector Machines on an audiovisual interest corpus (AVIC) consisting of spontaneous, conversational speech demonstrating ''theoretical'' effectiveness of the approach. Further, to evaluate the approach with regards to real-life usability a user-study is conducted for proof of ''practical'' effectiveness.