Supervised and semi-supervised infant-directed speech classification for parent-infant interaction analysis

  • Authors:
  • Ammar Mahdhaoui;Mohamed Chetouani

  • Affiliations:
  • Univ Paris 06, F-75005, Paris, France CNRS, UMR 7222, ISIR, Institut des Systèmes Intelligents et de Robotique, F-75005, Paris, France;Univ Paris 06, F-75005, Paris, France CNRS, UMR 7222, ISIR, Institut des Systèmes Intelligents et de Robotique, F-75005, Paris, France

  • Venue:
  • Speech Communication
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper describes the development of an infant-directed speech discrimination system for parent-infant interaction analysis. Different feature sets for emotion recognition were investigated using two classification techniques: supervised and semi-supervised. The classification experiments were carried out with short pre-segmented adult-directed speech and infant-directed speech segments extracted from real-life family home movies (with durations typically between 0.5s and 4s). The experimental results show that in the case of supervised learning, spectral features play a major role in the infant-directed speech discrimination. However, a major difficulty of using natural corpora is that the annotation process is time-consuming, and the expression of emotion is much more complex than in acted speech. Furthermore, interlabeler agreement and annotation label confidences are important issues to address. To overcome these problems, we propose a new semi-supervised approach based on the standard co-training algorithm exploiting labelled and unlabelled data. It offers a framework to take advantage of supervised classifiers trained by different features. The proposed dynamic weighted co-training approach combines various features and classifiers usually used in emotion recognition in order to learn from different views. Our experiments demonstrate the validity and effectiveness of this method for a real-life corpus such as home movies.