A scalable parallel algorithm for training a hierarchical mixture of neural experts

  • Authors:
  • Pablo A. Estévez;Hélène Paugam-Moisy;Didier Puzenat;Manuel Ugarte

  • Affiliations:
  • Departamento de Ingeniería Eléctrica, Universidad de Chile, Casilla 412-3, Santiago, Chile;Institut des Sciences Cognitives, UMR CNRS 5015, 67 boulevard Pinel, F-69675 Bron Cedex, France;Institut des Sciences Cognitives, UMR CNRS 5015, 67 boulevard Pinel, F-69675 Bron Cedex, France and Equipe GRIMAAG, Université Antilles-Guyane, Campus de Fouillole, F-97159 Pointe-à-Pitr ...;Departamento de Ingeniería Eléctrica, Universidad de Chile, Casilla 412-3, Santiago, Chile

  • Venue:
  • Parallel Computing
  • Year:
  • 2002

Quantified Score

Hi-index 0.00

Visualization

Abstract

Efficient parallel learning algorithms are proposed for training a powerful modular neural network, the hierarchical mixture of experts (HME). Parallelizations are based on the concept of modular parallelism, i.e. parallel execution of network modules. From modeling the speed-up as a function of the number of processors and the number of training examples, several improvements are derived, such as pipelining the training examples by packets. Compared to experimental measurements, theoretical models are accurate. For regular topologies, an analysis of the models shows that the parallel algorithms are highly scalable when the size of the experts grows from linear units to multi-layer perceptrons (MLPs). These results are confirmed experimentally, achieving near-linear speedups for HME-MLP. Although this work can be viewed as a case study in the parallelization of HME neural networks, both algorithms and theoretical models can be expanded to different learning rules or less regular tree architectures.