IKNN: Informative K-Nearest Neighbor Pattern Classification

  • Authors:
  • Yang Song;Jian Huang;Ding Zhou;Hongyuan Zha;C. Lee Giles

  • Affiliations:
  • Department of Computer Science and Engineering,;College of Information Sciences and Technology, The Pennsylvania State University, University Park, PA 16802, U.S.A.;Department of Computer Science and Engineering,;Department of Computer Science and Engineering, and College of Information Sciences and Technology, The Pennsylvania State University, University Park, PA 16802, U.S.A.;Department of Computer Science and Engineering, and College of Information Sciences and Technology, The Pennsylvania State University, University Park, PA 16802, U.S.A.

  • Venue:
  • PKDD 2007 Proceedings of the 11th European conference on Principles and Practice of Knowledge Discovery in Databases
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

The K-nearest neighbor (KNN) decision rule has been a ubiquitous classification tool with good scalability. Past experience has shown that the optimal choice of Kdepends upon the data, making it laborious to tune the parameter for different applications. We introduce a new metric that measures the informativeness of objects to be classified. When applied as a query-based distance metric to measure the closeness between objects, two novel KNN procedures, Locally Informative-KNN (LI-KNN) and Globally Informative-KNN (GI-KNN), are proposed. By selecting a subset of most informative objects from neighborhoods, our methods exhibit stability to the change of input parameters, number of neighbors(K) and informative points (I). Experiments on UCI benchmark data and diverse real-world data sets indicate that our approaches are application-independent and can generally outperform several popular KNN extensions, as well as SVM and Boosting methods.