The watershed transform: definitions, algorithms and parallelization strategies
Fundamenta Informaticae - Special issue on mathematical morphology
Robust automatic speech recognition with missing and unreliable acoustic data
Speech Communication
Monaural speech separation and recognition challenge
Computer Speech and Language
Data driven search organization for continuous speech recognition
IEEE Transactions on Signal Processing
Monaural speech separation and recognition challenge
Computer Speech and Language
The Markov selection model for concurrent speech recognition
Neurocomputing
Computer Speech and Language
International Journal of Speech Technology
Hi-index | 0.00 |
This paper addresses the problem of recognising speech in the presence of a competing speaker. We review a speech fragment decoding technique that treats segregation and recognition as coupled problems. Data-driven techniques are used to segment a spectro-temporal representation into a set of fragments, such that each fragment is dominated by one or other of the speech sources. A speech fragment decoder is used which employs missing data techniques and clean speech models to simultaneously search for the set of fragments and the word sequence that best matches the target speaker model. The paper investigates the performance of the system on a recognition task employing artificially mixed target and masker speech utterances. The fragment decoder produces significantly lower error rates than a conventional recogniser, and mimics the pattern of human performance that is produced by the interplay between energetic and informational masking. However, at around 0dB the performance is generally quite poor. An analysis of the errors shows that a large number of target/masker confusions are being made. The paper presents a novel fragment-based speaker identification approach that allows the target speaker to be reliably identified across a wide range of SNRs. This component is combined with the recognition system to produce significant improvements. When the target and masker utterance have the same gender, the recognition system has a performance at 0dB equal to that of humans; in other conditions the error rate is roughly twice the human error rate.