Expected browsing utility for web search evaluation

  • Authors:
  • Emine Yilmaz;Milad Shokouhi;Nick Craswell;Stephen Robertson

  • Affiliations:
  • Microsoft Research Cambridge, Cambridge, United Kingdom;Microsoft Research Cambridge, Cambridge, United Kingdom;Microsoft Research Cambridge, Cambridge, United Kingdom;Microsoft Research Cambridge, Cambridge, United Kingdom

  • Venue:
  • CIKM '10 Proceedings of the 19th ACM international conference on Information and knowledge management
  • Year:
  • 2010

Quantified Score

Hi-index 0.01

Visualization

Abstract

Most information retrieval evaluation metrics are designed to measure the satisfaction of the user given the results returned by a search engine. In order to evaluate user satisfaction, most of these metrics have underlying user models, which aim at modeling how users interact with search engine results. Hence, the quality of an evaluation metric is a direct function of the quality of its underlying user model. This paper proposes EBU, a new evaluation metric that uses a sophisticated user model tuned by observations over many thousands of real search sessions. We compare EBU with a number of state of the art evaluation metrics and show that it is more correlated with real user behavior captured by clicks.