Foundation of the DISIMA Image Query Languages

  • Authors:
  • Vincent Oria;M. Tamer Özsu;Paul J. Iglinski

  • Affiliations:
  • Department of Computer Science, New Jersey Institute of Technology. Vincent.Oria@njit.edu;School of Computer Science, University of Waterloo. tozsu@db.uwaterloo.ca;Department of Electrical & Computer Engineering, University of Alberta. iglinski@ee.ualberta.ca

  • Venue:
  • Multimedia Tools and Applications
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

Because digital images are not meaningful by themselves, images are often coupled with some descriptive or qualitative data in an image database. These data, divided into syntactic (color, shape, and texture) and semantic (meaningful real word object or concept) features, necessitate novel querying techniques. Most image systems and prototypes have focussed on similarity searches based upon the syntactic features. In the DISIMA system, we proposed an object-oriented image data model that introduces two main types: image (that represents an image and its descriptive properties) and salient object (that represents the semantics of an image). We further defined operations on the images and the salient objects as new joins. This approach is necessary in order to envision a declarative query language for images. This paper summarizes the querying facilities implemented for the DISIMA system and gives their theoretical foundation: the data model and the complementary algebraic operations, the textual query language (MOQL) and its visual counterpart (VisualMOQL) based on an image calculus. Both languages are declarative and allow the combination of semantic and similarity queries.