A new benchmark dataset with production methodology for short text semantic similarity algorithms

  • Authors:
  • James O'shea;Zuhair Bandar;Keeley Crockett

  • Affiliations:
  • Manchester Metropolitan University, Manchester, UK;Manchester Metropolitan University, Manchester, UK;Manchester Metropolitan University, Manchester, UK

  • Venue:
  • ACM Transactions on Speech and Language Processing (TSLP)
  • Year:
  • 2014

Quantified Score

Hi-index 0.00

Visualization

Abstract

This research presents a new benchmark dataset for evaluating Short Text Semantic Similarity (STSS) measurement algorithms and the methodology used for its creation. The power of the dataset is evaluated by using it to compare two established algorithms, STASIS and Latent Semantic Analysis. This dataset focuses on measures for use in Conversational Agents; other potential applications include email processing and data mining of social networks. Such applications involve integrating the STSS algorithm in a complex system, but STSS algorithms must be evaluated in their own right and compared with others for their effectiveness before systems integration. Semantic similarity is an artifact of human perception; therefore its evaluation is inherently empirical and requires benchmark datasets derived from human similarity ratings. The new dataset of 64 sentence pairs, STSS-131, has been designed to meet these requirements drawing on a range of resources from traditional grammar to cognitive neuroscience. The human ratings are obtained from a set of trials using new and improved experimental methods, with validated measures and statistics. The results illustrate the increased challenge and the potential longevity of the STSS-131 dataset as the Gold Standard for future STSS algorithm evaluation.