Bayesian inference of visual motion boundaries

  • Authors:
  • David J. Fleet;Michael J. Black;Oscar Nestares

  • Affiliations:
  • Palo Alto Research Center, 3333 Coyote Hill Road, Palo Alto, CA;Department of Computer Science, Brown University, Providence, RI;Instituto de Óptica (C.S.I.C.), Serrano 121, 28006-Madrid, Spain

  • Venue:
  • Exploring artificial intelligence in the new millennium
  • Year:
  • 2003

Quantified Score

Hi-index 0.03

Visualization

Abstract

This chapter addresses an open problem in visual motion analysis, the estimation of image motion in the vicinity of occlusion boundaries. With a Bayesian formulation, local image motion is explained in terms of multiple, competing, nonlinear models, including models for smooth (translational) motion and for motion boundaries. The generative model for motion boundaries explicitly encodes the orientation of the boundary, the velocities on either side, the motion of the occluding edge over time, and the appearance/disappearance of pixels at the boundary. We formulate the posterior probability distribution over the models and model parameters, conditioned on the image sequence. Approximate inference is achieved with a combination of tools: A Bayesian filter provides for online computation; factored sampling allows us to represent multimodal non-Gaussian distributions and to propagate beliefs with nonlinear dynamics from one time to the next; and mixture models are used to simplify the computation of joint prediction distributions in the Bayesian filter. To efficiently represent such a high-dimensional space, we also initialize samples using the responses of a low-level motion-discontinuity detector. The basic formulation and computational model provide a general probabilistic framework for motion estimation with multiple, nonlinear models.