Approximate nearest neighbors: towards removing the curse of dimensionality
STOC '98 Proceedings of the thirtieth annual ACM symposium on Theory of computing
An optimal algorithm for approximate nearest neighbor searching fixed dimensions
Journal of the ACM (JACM)
Normalized Cuts and Image Segmentation
IEEE Transactions on Pattern Analysis and Machine Intelligence
Multidimensional binary search trees used for associative searching
Communications of the ACM
Solving Standard Quadratic Optimization Problems via Linear, Semidefinite and Copositive Programming
Journal of Global Optimization
Complexity of finding dense subgraphs
Discrete Applied Mathematics
Better Prediction of Protein Cellular Localization Sites with the it k Nearest Neighbors Classifier
Proceedings of the 5th International Conference on Intelligent Systems for Molecular Biology
Indexing the Distance: An Efficient Method to KNN Processing
Proceedings of the 27th International Conference on Very Large Data Bases
Fast Nearest Neighbor Search in Medical Image Databases
VLDB '96 Proceedings of the 22th International Conference on Very Large Data Bases
Text Categorization Using Weight Adjusted k-Nearest Neighbor Classification
PAKDD '01 Proceedings of the 5th Pacific-Asia Conference on Knowledge Discovery and Data Mining
Point Matching under Large Image Deformations and Illumination Changes
IEEE Transactions on Pattern Analysis and Machine Intelligence
Kernel k-means: spectral clustering and normalized cuts
Proceedings of the tenth ACM SIGKDD international conference on Knowledge discovery and data mining
Robust moving least-squares fitting with sharp features
ACM SIGGRAPH 2005 Papers
Discovering large dense subgraphs in massive graphs
VLDB '05 Proceedings of the 31st international conference on Very large data bases
A Spectral Technique for Correspondence Problems Using Pairwise Constraints
ICCV '05 Proceedings of the Tenth IEEE International Conference on Computer Vision - Volume 2
Robust one-class clustering using hybrid global and local search
ICML '05 Proceedings of the 22nd international conference on Machine learning
Graphical Models and Point Pattern Matching
IEEE Transactions on Pattern Analysis and Machine Intelligence
Dominant Sets and Pairwise Clustering
IEEE Transactions on Pattern Analysis and Machine Intelligence
Shape Classification Using the Inner-Distance
IEEE Transactions on Pattern Analysis and Machine Intelligence
Voronoi-based K nearest neighbor search for spatial network databases
VLDB '04 Proceedings of the Thirtieth international conference on Very large data bases - Volume 30
A rate-distortion one-class model and its applications to clustering
Proceedings of the 25th international conference on Machine learning
Graph construction and b-matching for semi-supervised learning
ICML '09 Proceedings of the 26th Annual International Conference on Machine Learning
Fast Construction of k-Nearest Neighbor Graphs for Point Clouds
IEEE Transactions on Visualization and Computer Graphics
Semi-Supervised Learning
B-Matching for spectral clustering
ECML'06 Proceedings of the 17th European conference on Machine Learning
Hi-index | 0.00 |
In this paper, we study the problem of how to reliably compute neighborhoods on affinity graphs. The k-nearest neighbors (kNN) is one of the most fundamental and simple methods widely used in many tasks, such as classification and graph construction. Previous research focused on how to efficiently compute kNN on vectorial data. However, most real-world data have no vectorial representations, and only have affinity graphs which may contain unreliable affinities. Since the kNN of an object o is a set of k objects with the highest affinities to o, it is easily disturbed by errors in pairwise affinities between o and other objects, and also it cannot well preserve the structure underlying the data. To reliably analyze the neighborhood on affinity graphs, we define the k-dense neighborhood (kDN), which considers all pairwise affinities within the neighborhood, i.e., not only the affinities between o and its neighbors but also between the neighbors. For an object o, its kDN is a set kDN(o) of k objects which maximizes the sum of all pairwise affinities of objects in the set {o}驴kDN(o). We analyze the properties of kDN, and propose an efficient algorithm to compute it. Both theoretic analysis and experimental results on shape retrieval, semi-supervised learning, point set matching and data clustering show that kDN significantly outperforms kNN on affinity graphs, especially when many pairwise affinities are unreliable.