Latent word context model for information retrieval

  • Authors:
  • Bernard Brosseau-Villeneuve;Jian-Yun Nie;Noriko Kando

  • Affiliations:
  • University of Montréal, Montreal, Canada H3C 3J7;University of Montréal, Montreal, Canada H3C 3J7;National Institute of Informatics, Tokyo, Japan 101-8430

  • Venue:
  • Information Retrieval
  • Year:
  • 2014

Quantified Score

Hi-index 0.00

Visualization

Abstract

The application of word sense disambiguation (WSD) techniques to information retrieval (IR) has yet to provide convincing retrieval results. Major obstacles to effective WSD in IR include coverage and granularity problems of word sense inventories, sparsity of document context, and limited information provided by short queries. In this paper, to alleviate these issues, we propose the construction of latent context models for terms using latent Dirichlet allocation. We propose building one latent context per word, using a well principled representation of local context based on word features. In particular, context words are weighted using a decaying function according to their distance to the target word, which is learnt from data in an unsupervised manner. The resulting latent features are used to discriminate word contexts, so as to constrict query's semantic scope. Consistent and substantial improvements, including on difficult queries, are observed on TREC test collections, and the techniques combines well with blind relevance feedback. Compared to traditional topic modeling, WSD and positional indexing techniques, the proposed retrieval model is more effective and scales well on large-scale collections.