Auditory cortical representations of speech signals for phoneme classification

  • Authors:
  • Hugo L. Rufiner;César E. Martínez;Diego H. Milone;John Goddard

  • Affiliations:
  • Lab. de Señales e Inteligencia Computacional, Depto Informática, Facultad de Ingenieráa y Cs Hádricas, Univ. Nacional del Litoral, Santa Fe, Argentina and Fac. de Ingenierí ...;Lab. de Señales e Inteligencia Computacional, Depto Informática, Fac. de Ingenieráa y Cs Hádricas, Univ. Nacional del Litoral, Santa Fe, Argentina and Facultad de Ingenierí ...;Facultad de Ingeniería, Universidad Nacional de Entre Ríos, Argentina;Dpto. de Ingeniería Eléctrica, UAM, Iztapalapa, México

  • Venue:
  • MICAI'07 Proceedings of the artificial intelligence 6th Mexican international conference on Advances in artificial intelligence
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

The use of biologically inspired, feature extraction methods has improved the performance of artificial systems that try to emulate some aspect of human communication. Recent techniques, such as independent component analysis and sparse representations, have made it possible to undertake speech signal analysis using features similar to the ones found experimentally at the primary auditory cortex level. In this work, a new type of speech signal representation, based on the spectro-temporal receptive fields, is presented, and a problem of phoneme classification is tackled for the first time using this representation. The results obtained are compared, and found to greatly improve both an early auditory representation and the classical front-end based on Mel frequency cepstral coefficients.