PEM: a paraphrase evaluation metric exploiting parallel texts

  • Authors:
  • Chang Liu;Daniel Dahlmeier;Hwee Tou Ng

  • Affiliations:
  • National University of Singapore;NUS Graduate School for Integrative Sciences and Engineering;National University of Singapore and NUS Graduate School for Integrative Sciences and Engineering

  • Venue:
  • EMNLP '10 Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present PEM, the first fully automatic metric to evaluate the quality of paraphrases, and consequently, that of paraphrase generation systems. Our metric is based on three criteria: adequacy, fluency, and lexical dissimilarity. The key component in our metric is a robust and shallow semantic similarity measure based on pivot language N-grams that allows us to approximate adequacy independently of lexical similarity. Human evaluation shows that PEM achieves high correlation with human judgments.