Segmentation and Recognition Using Structure from Motion Point Clouds

  • Authors:
  • Gabriel J. Brostow;Jamie Shotton;Julien Fauqueur;Roberto Cipolla

  • Affiliations:
  • University College London and ETH Zurich,;Microsoft Research Cambridge,;University of Cambridge (now with MirriAd Ltd.),;University of Cambridge,

  • Venue:
  • ECCV '08 Proceedings of the 10th European Conference on Computer Vision: Part I
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

We propose an algorithm for semantic segmentation based on 3D point clouds derived from ego-motion. We motivate five simple cues designed to model specific patterns of motion and 3D world structure that vary with object category. We introduce features that project the 3D cues back to the 2D image plane while modeling spatial layout and context. A randomized decision forest combines many such features to achieve a coherent 2D segmentation and recognize the object categories present. Our main contribution is to show how semantic segmentation is possible based solely on motion-derived 3D world structure. Our method works well on sparse, noisy point clouds, and unlike existing approaches, does not need appearance-based descriptors.Experiments were performed on a challenging new video database containing sequences filmed from a moving car in daylight and at dusk. The results confirm that indeed, accurate segmentation and recognition are possible using only motion and 3D world structure. Further, we show that the motion-derived information complements an existing state-of-the-art appearance-based method, improving both qualitative and quantitative performance.