Sound classification in hearing aids inspired by auditory scene analysis

  • Authors:
  • Michael Büchler;Silvia Allegro;Stefan Launer;Norbert Dillier

  • Affiliations:
  • ENT Department, University Hospital Zurich, Zurich, Switzerland;Phonak AG, Staefa, Switzerland;Phonak AG, Staefa, Switzerland;ENT Department, University Hospital Zurich, Zurich, Switzerland

  • Venue:
  • EURASIP Journal on Applied Signal Processing
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

A sound classification system for the automatic recognition of the acoustic environment in a hearing aid is discussed. The system distinguishes the four sound classes "clean speech," "speech in noise," "noise," and "music." A number of features that are inspired by auditory scene analysis are extracted from the sound signal. These features describe amplitude modulations, spectral profile, harmonicity, amplitude onsets, and rhythm. They are evaluated together with different pattern classifiers. Simple classifiers, such as rule-based and minimum-distance classifiers, are compared with more complex approaches, such as Bayes classifier, neural network, and hidden Markov model. Sounds from a large database are employed for both training and testing of the system. The achieved recognition rates are very high except for the class "speech in noise." Problems arise in the classification of compressed pop music, strongly reverberated speech, and tonal or fluctuating noises.