Independent component analysis, a new concept?
Signal Processing - Special issue on higher order statistics
A fast fixed-point algorithm for independent component analysis
Neural Computation
EURASIP Journal on Applied Signal Processing
A binaural room impulse response database for the evaluation of dereverberation algorithms
DSP'09 Proceedings of the 16th international conference on Digital Signal Processing
Use of bimodal coherence to resolve spectral indeterminacy in Convolutive BSS
LVA/ICA'10 Proceedings of the 9th international conference on Latent variable analysis and signal separation
A blind source separation technique using second-order statistics
IEEE Transactions on Signal Processing
Matching pursuits with time-frequency dictionaries
IEEE Transactions on Signal Processing
IEEE Transactions on Audio, Speech, and Language Processing
IEEE Transactions on Audio, Speech, and Language Processing
IEEE Transactions on Audio, Speech, and Language Processing
Blind Audiovisual Source Separation Based on Sparse Redundant Representations
IEEE Transactions on Multimedia
IEEE Transactions on Neural Networks
Learning Bimodal Structure in Audio–Visual Data
IEEE Transactions on Neural Networks
Hi-index | 0.08 |
Recent studies show that facial information contained in visual speech can be helpful for the performance enhancement of audio-only blind source separation (BSS) algorithms. Such information is exploited through the statistical characterization of the coherence between the audio and visual speech using, e.g., a Gaussian mixture model (GMM). In this paper, we present three contributions. With the synchronized features, we propose an adapted expectation maximization (AEM) algorithm to model the audio-visual coherence in the off-line training process. To improve the accuracy of this coherence model, we use a frame selection scheme to discard nonstationary features. Then with the coherence maximization technique, we develop a new sorting method to solve the permutation problem in the frequency domain. We test our algorithm on a multimodal speech database composed of different combinations of vowels and consonants. The experimental results show that our proposed algorithm outperforms traditional audio-only BSS, which confirms the benefit of using visual speech to assist in separation of the audio.