On the geometry of feedforward neural network error surfaces

  • Authors:
  • An Mei Chen;Haw-minn Lu;Robert Hecht-Nielsen

  • Affiliations:
  • -;-;-

  • Venue:
  • Neural Computation
  • Year:
  • 1993

Quantified Score

Hi-index 0.00

Visualization

Abstract

Many feedforward neural network architectures have the propertythat their overall input-output function is unchanged by certainweight permutations and sign flips. In this paper, the geometricstructure of these equioutput weight space transformationsis explored for the case of multilayer perceptron networks withtanh activation functions (similar results hold for manyother types of neural networks). It is shown that thesetransformations form an algebraic group isomorphic to a directproduct of Weyl groups. Results concerning the root spaces of theLie algebras associated with these Weyl groups are then used toderive sets of simple equations for minimal sufficient search setsin weight space. These sets, which take the geometric forms of awedge and a cone, occupy only a minute fraction of the volume ofweight space. A separate analysis shows that large numbers ofcopies of a network performance function optimum weight vector arecreated by the action of the equioutput transformation group andthat these copies all lie on the same sphere. Some implications ofthese results for learning are discussed.