Towards a better understanding of language model information retrieval

  • Authors:
  • M. Van Der Heijden;I. G. Sprinkhuizen-Kuyper;Th. P. Van Der Weide

  • Affiliations:
  • Radboud University Nijmegen;Radboud University Nijmegen, Donders Institute for Brain Cognition and Behavior;Radboud University Nijmegen, Institute for Computing and Information Science

  • Venue:
  • FDIA'08 Proceedings of the 2nd BCS IRSG conference on Future Directions in Information Access
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

Language models form a class of successful probabilistic models in information retrieval. However, knowledge of why some methods perform better than others in a particular situation remains limited. In this study we analyze what language model factors influence information retrieval performance. Starting from popular smoothing methods we review what data features have been used. Document length and a measure of document word distribution turned out to be the important factors, in addition to a distinction in estimating the probability of seen and unseen words. We propose a class of parameter-free smoothing methods, of which multiple specific instances are possible. Instead of parameter tuning however, an analysis of data features should be used to decide upon a specific method. Finally, we discuss some initial experiments.