A probabilistic model for Latent Semantic Indexing: Research Articles

  • Authors:
  • Chris H. Q. Ding

  • Affiliations:
  • Lawrence Berkeley National Laboratory, University of California, Berkeley, CA 94720

  • Venue:
  • Journal of the American Society for Information Science and Technology
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

Latent Semantic Indexing (LSI), when applied to semantic space built on text collections, improves information retrieval, information filtering, and word sense disambiguation. A new dual probability model based on the similarity concepts is introduced to provide deeper understanding of LSI. Semantic associations can be quantitatively characterized by their statistical significance, the likelihood. Semantic dimensions containing redundant and noisy information can be separated out and should be ignored because their negative contribution to the overall statistical significance. LSI is the optimal solution of the model. The peak in the likelihood curve indicates the existence of an intrinsic semantic dimension. The importance of LSI dimensions follows the Zipf-distribution, indicating that LSI dimensions represent latent concepts. Document frequency of words follows the Zipf distribution, and the number of distinct words follows log-normal distribution. Experiments on five standard document collections confirm and illustrate the analysis. © 2005 Wiley Periodicals, Inc.