Learning the Latent Semantic Space for Ranking in Text Retrieval

  • Authors:
  • Jun Yan;Shuicheng Yan;Ning Liu;Zheng Chen

  • Affiliations:
  • -;-;-;-

  • Venue:
  • ICDM '08 Proceedings of the 2008 Eighth IEEE International Conference on Data Mining
  • Year:
  • 2008

Quantified Score

Hi-index 0.02

Visualization

Abstract

Subspace learning techniques for text analysis, such as Latent Semantic Indexing (LSI), have been widely studied in the past decade. However, to our best knowledge, no previous study has leveraged the rank information for subspace learning in ranking tasks. In this paper, we propose a novel algorithm, called Learning Latent Semantics for Ranking (LLSR), to seek the optimal Latent Semantic Space tailored to the ranking tasks. We first present a dual explanation for the classical Latent Semantic Indexing (LSI) algorithm, namely learning the so-called Latent Semantic Space (LSS) to encode the data information. Then, to handle the increasing amount of training data for the practical ranking tasks, we propose a novel objective function to derive the optimal LSS for ranking. Experimental results on two SMART sub-collections and a TREC dataset show that LLSR effectively improves the ranking performance compared with the classical LSI algorithm and ranking without subspace learning.