Computing Optimal Attribute Weight Settings for Nearest NeighborAlgorithms

  • Authors:
  • Charles X. Ling;Hangdong Wang

  • Affiliations:
  • Department of Computer Science, M.C., The University of Western Ontario, London, Ontario, Canada N6A 5B7. E-mail: ling@csd.uwo.ca, hwang@csd.uwo.ca;Department of Computer Science, M.C., The University of Western Ontario, London, Ontario, Canada N6A 5B7. E-mail: ling@csd.uwo.ca, hwang@csd.uwo.ca

  • Venue:
  • Artificial Intelligence Review - Special issue on lazy learning
  • Year:
  • 1997

Quantified Score

Hi-index 0.00

Visualization

Abstract

Nearest neighbor (NN) learning algorithms, examples of the lazy learningparadigm, rely on a distance function to measure the similarity of testingexamples with the stored training examples. Since certain attributes are morediscriminative, while others can be less or totally irrelevant, attributesshould be weighed differently in the distance function. Most previous studieson weight setting for NN learning algorithms are empirical. In this paper wedescribe our attempt on deciding theoretically optimal weights that minimizethe predictive error for NN algorithms. Assuming a uniform distribution ofexamples in a 2-d continuous space, we first derive the average predictiveerror introduced by a linear classification boundary, and then determine theoptimal weight setting for any polygonal classification region. Our theoreticalresults of optimal attribute weights can serve as a baseline or lower bound forcomparing other empirical weight setting methods.