Experimental study on prototype optimisation algorithms for prototype-based classification in vector spaces

  • Authors:
  • M. Lozano;J. M. Sotoca;J. S. Sánchez;F. Pla;E. Pkalska;R. P. W. Duin

  • Affiliations:
  • Dept. Lenguajes y Sistemas Informáticos, Universitat Jaume I, Campus Riu Sec, 12071 Castellón, Spain;Dept. Lenguajes y Sistemas Informáticos, Universitat Jaume I, Campus Riu Sec, 12071 Castellón, Spain;Dept. Lenguajes y Sistemas Informáticos, Universitat Jaume I, Campus Riu Sec, 12071 Castellón, Spain;Dept. Lenguajes y Sistemas Informáticos, Universitat Jaume I, Campus Riu Sec, 12071 Castellón, Spain;Faculty of Electrical Engineering, Mathematics and Computer Science, Delft University of Technology, Mekelweg 4, 2628 CD Delft, The Netherlands and School of Computer Science, The University of Ma ...;Faculty of Electrical Engineering, Mathematics and Computer Science, Delft University of Technology, Mekelweg 4, 2628 CD Delft, The Netherlands

  • Venue:
  • Pattern Recognition
  • Year:
  • 2006

Quantified Score

Hi-index 0.01

Visualization

Abstract

Prototype-based classification relies on the distances between the examples to be classified and carefully chosen prototypes. A small set of prototypes is of interest to keep the computational complexity low, while maintaining high classification accuracy. An experimental study of some old and new prototype optimisation techniques is presented, in which the prototypes are either selected or generated from the given data. These condensing techniques are evaluated on real data, represented in vector spaces, by comparing their resulting reduction rates and classification performance. Usually the determination of prototypes is studied in relation with the nearest neighbour rule. We will show that the use of more general dissimilarity-based classifiers can be more beneficial. An important point in our study is that the adaptive condensing schemes here discussed allow the user to choose the number of prototypes freely according to the needs. If such techniques are combined with linear dissimilarity-based classifiers, they provide the best trade-off of small condensed sets and high classification accuracy.