Shared-task evaluations in HLT: lessons for NLG

  • Authors:
  • Anja Belz;Adam Kilgarriff

  • Affiliations:
  • University of Brighton, UK;Lexical Computing Ltd., UK

  • Venue:
  • INLG '06 Proceedings of the Fourth International Natural Language Generation Conference
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

While natural language generation (NLG) has a strong evaluation tradition, in particular in userbased and task-oriented evaluation, it has never evaluated different approaches and techniques by comparing their performance on the same tasks (shared-task evaluation, STE). NLG is characterised by a lack of consolidation of results, and by isolation from the rest of NLP where STE is now standard. It is, moreover, a shrinking field (state-of-the-art MT and summarisation no longer perform generation as a subtask) which lacks the kind of funding and participation that natural language understanding (NLU) has attracted.