Efficient Nearest-Neighbor Search in the Probability Simplex

  • Authors:
  • Kriste Krstovski;David A. Smith;Hanna M. Wallach;Andrew McGregor

  • Affiliations:
  • School of Computer Science, University of Massachusetts, Amherst, MA, 01003, U.S.A.;College of Computer and Information Science, Northeastern University, Boston, MA, 02115, U.S.A.;School of Computer Science, University of Massachusetts, Amherst, MA, 01003, U.S.A.;School of Computer Science, University of Massachusetts, Amherst, MA, 01003, U.S.A.

  • Venue:
  • Proceedings of the 2013 Conference on the Theory of Information Retrieval
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Document similarity tasks arise in many areas of information retrieval and natural language processing. A fundamental question when comparing documents is which representation to use. Topic models, which have served as versatile tools for exploratory data analysis and visualization, represent documents as probability distributions over latent topics. Systems comparing topic distributions thus use measures of probability divergence such as Kullback-Leibler, Jensen-Shannon, or Hellinger. This paper presents novel analysis and applications of the reduction of Hellinger divergence to Euclidean distance computations. This reduction allows us to exploit fast approximate nearest-neighbor (NN) techniques, such as locality-sensitive hashing (LSH) and approximate search in k-d trees, for search in the probability simplex. We demonstrate the effectiveness and efficiency of this approach on two tasks using latent Dirichlet allocation (LDA) document representations: discovering relationships between National Institutes of Health (NIH) grants and prior-art retrieval for patents. Evaluation on these tasks and on synthetic data shows that both Euclidean LSH and approximate k-d tree search perform well when a single nearest neighbor must be found. When a larger set of similar documents is to be retrieved, the k-d tree approach is more effective and efficient.