Artificial Intelligence Review - Special issue on lazy learning
Statistical Pattern Recognition: A Review
IEEE Transactions on Pattern Analysis and Machine Intelligence
Machine Learning
A Tutorial on Support Vector Machines for Pattern Recognition
Data Mining and Knowledge Discovery
Vector Quantization Technique for Nonparametric Classifier Design
IEEE Transactions on Pattern Analysis and Machine Intelligence
Collaborative recommendation: A robustness analysis
ACM Transactions on Internet Technology (TOIT)
IEEE Transactions on Pattern Analysis and Machine Intelligence
A clustering-based method for unsupervised intrusion detections
Pattern Recognition Letters
Face recognition based on discriminant fractional Fourier feature extraction
Pattern Recognition Letters
Finding Prototypes For Nearest Neighbor Classifiers
IEEE Transactions on Computers
Locally linear reconstruction for instance-based learning
Pattern Recognition
IEEE Transactions on Pattern Analysis and Machine Intelligence
Enhancing prototype reduction schemes with recursion: a method applicable for "large" data sets
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
The condensed nearest neighbor rule (Corresp.)
IEEE Transactions on Information Theory
An algorithm for a selective nearest neighbor decision rule (Corresp.)
IEEE Transactions on Information Theory
Hi-index | 0.01 |
This paper concerns the use of prototype reduction schemes (PRS) to optimize the computations involved in typical k-nearest neighbor (k-NN) rules. These rules have been successfully used for decades in statistical pattern recognition (PR) [1,15] applications and are particularly effective for density estimation, classification, and regression because of the known error bounds that they possess. For a given data point of unknown identity, the k-NN possesses the phenomenon that it combines the information about the samples from a priori target classes (values) of selected neighbors to predict the target class of the tested sample, or to estimate the density function value of the given queried sample. Recently, an implementation of the k-NN, named as the locally linear reconstruction (LLR) [2], has been proposed. The salient and brilliant feature of the latter is that by invoking a quadratic optimization process, it is capable of systematically setting model parameters, such as the number of neighbors (specified by the parameter, k) and the weights. However, the LLR takes more time than other conventional methods when it has to be applied to classification tasks. To overcome this problem, we propose a strategy of using a PRS to efficiently compute the optimization problem. In this paper, we demonstrate, first of all, that by completely discarding the points not included by the PRS, we can obtain a reduced set of sample points, using which, in turn, the quadratic optimization problem can be computed far more expediently. The values of the corresponding indices are comparable to those obtained with the original training set (i.e., the one which considers all the data points) even though the computations required to obtain the prototypes and the corresponding classification accuracies are noticeably less. The proposed method has been tested on artificial and real-life data sets, and the results obtained are very promising, and could have potential in PR applications.