Evolutionary training of hardware realizable multilayer perceptrons

  • Authors:
  • P. Plagianakos;D. Magoulas;N. Vrahatis

  • Affiliations:
  • Computational Intelligence Laboratory, Department of Mathematics, University of Patras Artificial Intelligence Research Center–UPAIRC, University of Patras, GR-26110, Patras, Greece;School of Computer Science and Information Systems, University of London, Malet Street, WC1E7HX, London, UK;Computational Intelligence Laboratory, Department of Mathematics, University of Patras Artificial Intelligence Research Center–UPAIRC, University of Patras, GR-26110, Patras, Greece

  • Venue:
  • Neural Computing and Applications
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

The use of multilayer perceptrons (MLP) with threshold functions (binary step function activations) greatly reduces the complexity of the hardware implementation of neural networks, provides tolerance to noise and improves the interpretation of the internal representations. In certain case, such as in learning stationary tasks, it may be sufficient to find appropriate weights for an MLP with threshold activation functions by software simulation and, then, transfer the weight values to the hardware implementation. Efficient training of these networks is a subject of considerable ongoing research. Methods available in the literature mainly focus on two-state (threshold) nodes and try to train the networks by approximating the gradient of the error function and modifying appropriately the gradient descent, or by progressively altering the shape of the activation functions. In this paper, we propose an evolution-motivated approach, which is eminently suitable for networks with threshold functions and compare its performance with four other methods. The proposed evolutionary strategy does not need gradient related information, it is applicable to a situation where threshold activations are used from the beginning of the training, as in “on-chip” training, and is able to train networks with integer weights.