Independent component analysis, a new concept?
Signal Processing - Special issue on higher order statistics
Natural gradient works efficiently in learning
Neural Computation
Blind separation methods based on Pearson system and its extensions
Signal Processing
Beyond Comon's Identifiability Theorem for Independent Component Analysis
ICANN '02 Proceedings of the International Conference on Artificial Neural Networks
Independent component analysis based on nonparametric density estimation
IEEE Transactions on Neural Networks
Hi-index | 0.03 |
Data processing and source identification using lower dimensional hidden structure plays an essential role in many fields of applications, including image processing, neural networks, genome studies, signal processing and other areas where large datasets are often encountered. One of the common methods for source separation using lower dimensional structure involves the use of Independent Component Analysis, which is based on a linear representation of the observed data in terms of independent hidden sources. The problem thus involves the estimation of the linear mixing matrix and the densities of the independent hidden sources. However, the solution to the problem depends on the identifiability of the sources. This paper first presents a set of sufficient conditions to establish the identifiability of the sources and the mixing matrix using moment restrictions of the hidden source variables. Under such sufficient conditions a semi-parametric maximum likelihood estimate of the mixing matrix is obtained using a class of mixture distributions. The consistency of our proposed estimate is established under additional regularity conditions. The proposed method is illustrated and compared with existing methods using simulated and real datasets.