Vision-based global localization for mobile robots with hybrid maps of objects and spatial layouts

  • Authors:
  • Soonyong Park;Soohwan Kim;Mignon Park;Sung-Kee Park

  • Affiliations:
  • Center for Cognitive Robotics Research, Korea Institute of Science and Technology (KIST), 39-1 Hawolgok-dong, Sungbuk-gu, Seoul, Republic of Korea;Center for Cognitive Robotics Research, Korea Institute of Science and Technology (KIST), 39-1 Hawolgok-dong, Sungbuk-gu, Seoul, Republic of Korea;Department of Electrical and Electronic Engineering, Yonsei University, 134 Sinchon-dong, Seodaemun-gu, Seoul, Republic of Korea;Center for Cognitive Robotics Research, Korea Institute of Science and Technology (KIST), 39-1 Hawolgok-dong, Sungbuk-gu, Seoul, Republic of Korea

  • Venue:
  • Information Sciences: an International Journal
  • Year:
  • 2009

Quantified Score

Hi-index 0.07

Visualization

Abstract

This paper presents a novel vision-based global localization that uses hybrid maps of objects and spatial layouts. We model indoor environments with a stereo camera using the following visual cues: local invariant features for object recognition and their 3D positions for object pose estimation. We also use the depth information at the horizontal centerline of image where the optical axis passes through, which is similar to the data from a 2D laser range finder. This allows us to build our topological node that is composed of a horizontal depth map and an object location map. The horizontal depth map describes the explicit spatial layout of each local space and provides metric information to compute the spatial relationships between adjacent spaces, while the object location map contains the pose information of objects found in each local space and the visual features for object recognition. Based on this map representation, we suggest a coarse-to-fine strategy for global localization. The coarse pose is estimated by means of object recognition and SVD-based point cloud fitting, and then is refined by stochastic scan matching. Experimental results show that our approaches can be used for an effective vision-based map representation as well as for global localization methods.