Approximate nearest neighbors: towards removing the curse of dimensionality
STOC '98 Proceedings of the thirtieth annual ACM symposium on Theory of computing
Similarity estimation techniques from rounding algorithms
STOC '02 Proceedings of the thiry-fourth annual ACM symposium on Theory of computing
Multiclass Spectral Clustering
ICCV '03 Proceedings of the Ninth IEEE International Conference on Computer Vision - Volume 2
Prototype vector machine for large scale semi-supervised learning
ICML '09 Proceedings of the 26th Annual International Conference on Machine Learning
Self-taught hashing for fast similarity search
Proceedings of the 33rd international ACM SIGIR conference on Research and development in information retrieval
Product Quantization for Nearest Neighbor Search
IEEE Transactions on Pattern Analysis and Machine Intelligence
Error-correcting output hashing in fast similarity search
ICIMCS '10 Proceedings of the Second International Conference on Internet Multimedia Computing and Service
Efficient manifold ranking for image retrieval
Proceedings of the 34th international ACM SIGIR conference on Research and development in Information Retrieval
Iterative quantization: A procrustean approach to learning binary codes
CVPR '11 Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition
Compact hashing with joint optimization of search accuracy and time
CVPR '11 Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition
Fast approximate nearest-neighbor search with k-nearest neighbor graph
IJCAI'11 Proceedings of the Twenty-Second international joint conference on Artificial Intelligence - Volume Volume Two
Learning hash functions for cross-view similarity search
IJCAI'11 Proceedings of the Twenty-Second international joint conference on Artificial Intelligence - Volume Volume Two
Semantic context learning with large-scale weakly-labeled image set
Proceedings of the 21st ACM international conference on Information and knowledge management
Hi-index | 0.00 |
Hashing-based fast nearest neighbor search technique has attracted great attention in both research and industry areas recently. Many existing hashing approaches encode data with projection-based hash functions and represent each projected dimension by 1-bit. However, the dimensions with high variance hold large energy or information of data but treated equivalently as dimensions with low variance, which leads to a serious information loss. In this paper, we introduce a novel hashing algorithm called Harmonious Hashing which aims at learning hash functions with low information loss. Specifically, we learn a set of optimized projections to preserve the maximum cumulative energy and meet the constraint of equivalent variance on each dimension as much as possible. In this way, we could minimize the information loss after binarization. Despite the extreme simplicity, our method outperforms superiorly to many state-of-the-art hashing methods in large-scale and high-dimensional nearest neighbor search experiments.