Self-supervised learning to visually detect terrain surfaces for autonomous robots operating in forested terrain

  • Authors:
  • Shengyan Zhou;Junqiang Xi;Matthew W. McDaniel;Takayuki Nishihata;Phil Salesses;Karl Iagnemma

  • Affiliations:
  • Department of Mechanical Engineering, Beijing Institute of Technology, Haidian, Beijing 100081, China;Department of Mechanical Engineering, Beijing Institute of Technology, Haidian, Beijing 100081, China;Robotic Mobility Group, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139;Robotic Mobility Group, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139;United States Army, Topographic Engineering Center, Alexandria, Virginia, 22315;Robotic Mobility Group, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139

  • Venue:
  • Journal of Field Robotics
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

Autonomous robotic navigation in forested environments is difficult because of the highly variable appearance and geometric properties of the terrain. In most navigation systems, researchers assume a priori knowledge of the terrain appearance properties, geometric properties, or both. In forest environments, vegetation such as trees, shrubs, and bushes has appearance and geometric properties that vary with change of seasons, vegetation age, and vegetation species. In addition, in forested environments the terrain surface is often rough, sloped, and/or covered with a surface layer of grass, vegetation, or snow. The complexity of the forest environment presents difficult challenges for autonomous navigation systems. In this paper, a self-supervised sensing approach is introduced that attempts to robustly identify a drivable terrain surface for robots operating in forested terrain. The sensing system employs both LIDAR and vision sensor data. There are three main stages in the system: feature learning, feature training, and terrain prediction. In the feature learning stage, 3D range points from LIDAR are analyzed to obtain an estimate of the ground surface location. In the feature training stage, the ground surface estimate is used to train a visual classifier to discriminate between ground and nonground regions of the image. In the prediction stage, the ground surface location can be estimated at high frequency solely from vision sensor data. © 2012 Wiley Periodicals, Inc. © 2012 Wiley Periodicals, Inc.