Comparison of methods for language-dependent and language-independent query-by-example spoken term detection

  • Authors:
  • Javier Tejedor;Michal Fapšo;Igor Szöke;Jan “Honza” Černocký;František Grézl

  • Affiliations:
  • Universidad Aut´onoma de Madrid, Spain;Brno University of Technology, Czech Republic;Brno University of Technology, Czech Republic;Brno University of Technology, Czech Republic;Brno University of Technology, Czech Republic

  • Venue:
  • ACM Transactions on Information Systems (TOIS)
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

This article investigates query-by-example (QbE) spoken term detection (STD), in which the query is not entered as text, but selected in speech data or spoken. Two feature extractors based on neural networks (NN) are introduced: the first producing phone-state posteriors and the second making use of a compressive NN layer. They are combined with three different QbE detectors: while the Gaussian mixture model/hidden Markov model (GMM/HMM) and dynamic time warping (DTW) both work on continuous feature vectors, the third one, based on weighted finite-state transducers (WFST), processes phone lattices. QbE STD is compared to two standard STD systems with text queries: acoustic keyword spotting and WFST-based search of phone strings in phone lattices. The results are reported on four languages (Czech, English, Hungarian, and Levantine Arabic) using standard metrics: equal error rate (EER) and two versions of popular figure-of-merit (FOM). Language-dependent and language-independent cases are investigated; the latter being particularly interesting for scenarios lacking standard resources to train speech recognition systems. While the DTW and GMM/HMM approaches produce the best results for a language-dependent setup depending on the target language, the GMM/HMM approach performs the best dealing with a language-independent setup. As far as WFSTs are concerned, they are promising as they allow for indexing and fast search.