Hierarchical mixtures of experts and the EM algorithm
Neural Computation
Optimal linear combinations of neural networks
Optimal linear combinations of neural networks
From data mining to knowledge discovery: an overview
Advances in knowledge discovery and data mining
IEEE Transactions on Pattern Analysis and Machine Intelligence
Adaptive mixtures of local experts
Neural Computation
Hi-index | 0.00 |
Learning problem has three distinct phases, that is, model representation, learning criterion (target function) and implementation algorithm. This paper focuses on the close relation between the selection of learning criterion for committee machine and network approximation and competitive adaptation. By minimizing the KL deviation between posterior distributions, we give a general posterior modular architecture and the corresponding learning criterion form, which reflects remarkable adaptation and scalability. Besides this, we point out, from the generalized KL deviation defined on finite measure manifold in information geometry theory, that the proposed learning criterion reduces to so-called Mahalanobis deviation of which ordinary mean square error approximation is a special case, when each module is assumed Gaussian.