From Images to Sentences via Spatial Relations

  • Authors:
  • Alicia Abella;John R. Kender

  • Affiliations:
  • -;-

  • Venue:
  • SPELMG '99 Proceedings of the Integration of Speech and Image Understanding
  • Year:
  • 1999

Quantified Score

Hi-index 0.00

Visualization

Abstract

This work presents a conceptual framework for representing, manipulating, measuring, and communicating in natural language several ideas about topological (non-metric) spatial locations, object spatial contexts, and user expectations of spatial relationships. It articulates a theory of spatial relations, how they can be represented as fuzzy predicates internally, and how they can be appropriately derived from imagery; then, how they can be augmented or filtered using prior knowledge; and lastly, how they can produce natural language statements about location and space. This framework quantifies the notions of context and vagueness, so that all spatial relations are measurably accurate, provably efficient, and matched to users' expectations.The work makes explicit two critical heuristics for reducing the complexity of the relationships implicit in imagery, one a general rule for single object descriptions, and the other a general rule for rank ordering object relationships.A derived working system combines variable aspects of computer science and linguistics in such a way so as to be extensible to many environments. The system has been demonstrated both in a landmark navigation task and in a medical task, two very separate domains, and has been evaluated in both.