Reshaping automatic speech transcripts for robust high-level spoken document analysis

  • Authors:
  • Julien Fayolle;Fabienne Moreau;Christian Raymond;Guillaume Gravier

  • Affiliations:
  • INRIA - IRISA, Rennes, France;University Rennes 2 - IRISA, Rennes, France;INSA - IRISA, Rennes, France;CNRS - IRISA, Rennes, France

  • Venue:
  • AND '10 Proceedings of the fourth workshop on Analytics for noisy unstructured text data
  • Year:
  • 2010

Quantified Score

Hi-index 0.01

Visualization

Abstract

High-level spoken document analysis is required in many applications seeking access to the semantic content of audio data, such as information retrieval, machine translation or automatic summarization. It is nevertheless a difficult task that is generally based on transcripts provided by an automatic speech recognition system. Unlike standard texts, transcripts belong to the category of highly noisy data because of word recognition errors that affect, in particular, very significant words such as named entities (e.g. person's names, locations, organizations). Transcripts also contain specificities of spoken language that make ineffective their processing by natural language processing tools designed for texts. To overcome these issues, this paper proposes a method to reshape automatic speech transcripts for robust high-level spoken document analysis. The method consists in conceiving a new word-level confidence measure that may efficiently ensure the reliability of transcribed words, focusing on words that are relevant for high-level spoken document analysis such as named entities. The approach consists in combining different features collected from various sources of knowledge thanks to a machine learning method based on conditional random fields. In addition to standard features (morphosyntactic, linguistic and phonetic), we introduce new semantic features based on the decisions of three robust named entity recognition systems to better estimate the reliability of named entities. Experiments, conducted on the French broadcast news corpus ESTER, demonstrate the added-value of the proposed word-level confidence measure for error detection and named entity recognition, with respect to the basic confidence measure provided by an automatic speech recognition system.