Towards an architecture for cognitive vision using qualitative spatio-temporal representations and abduction

  • Authors:
  • Anthony G. Cohn;Derek R. Magee;Aphrodite Galata;David C. Hogg;Shyamanta M. Hazarika

  • Affiliations:
  • School of Computing, University of Leeds, Leeds, UK;School of Computing, University of Leeds, Leeds, UK;School of Computing, University of Leeds, Leeds, UK;School of Computing, University of Leeds, Leeds, UK;School of Computing, University of Leeds, Leeds, UK

  • Venue:
  • Spatial cognition III
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

In recent years there has been increasing interest in constructing cognitive vision systems capable of interpreting the high level semantics of dynamic scenes. Purely quantitative approaches to the task of constructing such systems have met with some success. However, qualitative analysis of dynamic scenes has the advantage of allowing easier generalisation of classes of different behaviours and guarding against the propagation of errors caused by uncertainty and noise in the quantitative data. Our aim is to integrate quantitative and qualitative modes of representation and reasoning for the analysis of dynamic scenes. In particular, in this paper we outline an approach for constructing cognitive vision systems using qualitative spatial-temporal representations including prototypical spatial relations and spatio-temporal event descriptors automatically inferred from input data. The overall architecture relies on abduction: the system searches for explanations, phrased in terms of the learned spatio-temporal event descriptors, to account for the video data.