Matching utterances to rich knowledge structures to acquire a model of the speaker's goal

  • Authors:
  • Peter Z. Yeh;Bruce Porter;Ken Barker

  • Affiliations:
  • University of Texas, Austin, Austin, TX;University of Texas, Austin, Austin, TX;University of Texas, Austin, Austin, TX

  • Venue:
  • Proceedings of the 3rd international conference on Knowledge capture
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

An ultimate goal of AI is to build end-to-end systems that interpret natural language, reason over the resulting logical forms, and perform actions based on that reasoning. This requires systems from separate fields be brought together, but often this exposes representational gaps between them. The logical forms from a language interpreter may mirror the surface forms of utterances too closely to be usable as-is, given a reasoner's requirements for knowledge representations. What is needed is a system that can match logical forms to background knowledge flexibly to acquire a rich semantic model of the speaker's goal. In this paper, we present such a "matcher" that uses semantic transformations to overcome structural differences between the two representations. We evaluate this matcher in a MUC-like template-filling task and compare its performance to that of two similar systems.