Occlusion handling in video segmentation via predictive feedback

  • Authors:
  • Jeremie Papon;Alexey Abramov;Florentin Wörgötter

  • Affiliations:
  • Bernstein Center for Computational Neuroscience (BCCN), III Physikalisches Institut - Biophysik, Georg-August University of Göttingen, Germany;Bernstein Center for Computational Neuroscience (BCCN), III Physikalisches Institut - Biophysik, Georg-August University of Göttingen, Germany;Bernstein Center for Computational Neuroscience (BCCN), III Physikalisches Institut - Biophysik, Georg-August University of Göttingen, Germany

  • Venue:
  • ECCV'12 Proceedings of the 12th international conference on Computer Vision - Volume Part III
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present a method for unsupervised on-line dense video segmentation which utilizes sequential Bayesian estimation techniques to resolve partial and full occlusions. Consistent labeling through occlusions is vital for applications which move from low-level object labels to high-level semantic knowledge - tasks such as activity recognition or robot control. The proposed method forms a predictive loop between segmentation and tracking, with tracking predictions used to seed the segmentation kernel, and segmentation results used to update tracked models. All segmented labels are tracked, without the use of a-priori models, using parallel color-histogram particle filters. Predictions are combined into a probabilistic representation of image labels, a realization of which is used to seed segmentation. A simulated annealing relaxation process allows the realization to converge to a minimal energy segmented image. Found segments are subsequently used to repopulate the particle sets, closing the loop. Results on the Cranfield benchmark sequence demonstrate that the prediction mechanism allows on-line segmentation to maintain temporally consistent labels through partial & full occlusions, significant appearance changes, and rapid erratic movements. Additionally, we show that tracking performance matches state-of-the art tracking methods on several challenging benchmark sequences.