Generalized relevance learning vector quantization

  • Authors:
  • Barbara Hammer;Thomas Villmann

  • Affiliations:
  • Department of Mathematics and Computer Science, University of Osnabrück, Albrechtstraße 28, 49069 Osnabrück, Germany;Clinic for Psychotherapy and Psychosomatic Medicine, University of Leipzig, Karl-Tauchnitz-Straße 25, 04107 Leipzig, Germany

  • Venue:
  • Neural Networks - New developments in self-organizing maps
  • Year:
  • 2002

Quantified Score

Hi-index 0.00

Visualization

Abstract

We propose a new scheme for enlarging generalized learning vector quantization (GLVQ) with weighting factors for the input dimensions. The factors allow an appropriate scaling of the input dimensions according to their relevance. They are adapted automatically during training according to the specific classification task whereby training can be interpreted as stochastic gradient descent on an appropriate error function. This method leads to a more powerful classifier and to an adaptive metric with little extra cost compared to standard GLVQ. Moreover, the size of the weighting factors indicates the relevance of the input dimensions. This proposes a scheme for automatically pruning irrelevant input dimensions. The algorithm is verified on artificial data sets and the iris data from the UCI repository. Afterwards, the method is compared to several well known algorithms which determine the intrinsic data dimension on real world satellite image data.