A robust framework for joint background/foreground segmentation of complex video scenes filmed with freely moving camera

  • Authors:
  • Slim Amri;Walid Barhoumi;Ezzeddine Zagrouba

  • Affiliations:
  • Equipe de Recherche Systèmes Intelligents en Imagerie et Vision Artificielle (SIIVA) Institut Supérieur d'Informatique, Ariana, Tunisia 2080;Equipe de Recherche Systèmes Intelligents en Imagerie et Vision Artificielle (SIIVA) Institut Supérieur d'Informatique, Ariana, Tunisia 2080;Equipe de Recherche Systèmes Intelligents en Imagerie et Vision Artificielle (SIIVA) Institut Supérieur d'Informatique, Ariana, Tunisia 2080

  • Venue:
  • Multimedia Tools and Applications
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper explores a robust region-based general framework for discriminating between background and foreground objects within a complex video sequence. The proposed framework works under difficult conditions such as dynamic background and nominally moving camera. The originality of this work lies essentially in our use of the semantic information provided by the regions while simultaneously identifying novel objects (foreground) and non-novel ones (background). The information of background regions is exploited to make moving objects detection more efficient, and vice-versa. In fact, an initial panoramic background is modeled using region-based mosaicing in order to be sufficiently robust to noise from lighting effects and shadowing by foreground objects. After the elimination of the camera movement using motion compensation, the resulting panoramic image should essentially contain the background and the ghost-like traces of the moving objects. Then, while comparing the panoramic image of the background with the individual frames, a simple median-based background subtraction permits a rough identification of foreground objects. Joint background-foreground validation, based on region segmentation, is then used for a further examination of individual foreground pixels intended to eliminate false positives and to localize shadow effects. Thus, we first obtain a foreground mask from a slow-adapting algorithm, and then validate foreground pixels (moving visual objects + shadows) by a simple moving object model built by using both background and foreground regions. The tests realized on various well-known challenging real videos (across a variety of domains) show clearly the robustness of the suggested solution. This solution, which is relatively computationally inexpensive, can be used under difficult conditions such as dynamic background, nominally moving camera and shadows. In addition to the visual evaluation, spatial-based evaluation statistics, given hand-labeled ground truth, has been used as a performance measure of moving visual objects detection.