Reuse and challenges in evaluating language generation systems: position paper

  • Authors:
  • Kalina Bontcheva

  • Affiliations:
  • University of Sheffield, Sheffield, UK

  • Venue:
  • Evalinitiatives '03 Proceedings of the EACL 2003 Workshop on Evaluation Initiatives in Natural Language Processing: are evaluation methods, metrics and resources reusable?
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

Although there is an increasing shift towards evaluating Natural Language Generation (NLG) systems, there are still many NLG-specific open issues that hinder effective comparative and quantitative evaluation in this field. The paper starts off by describing a task-based, i.e., black-box evaluation of a hypertext NLG system. Then we examine the problem of glass-box, i.e., module specific, evaluation in language generation, with focus on evaluating machine learning methods for text planning.