Finding Text in Natural Scenes by Figure-Ground Segmentation

  • Authors:
  • Huiying Shen;James Coughlan

  • Affiliations:
  • Smith-Kettlewell Eye Research Institute, San Francisco, CA 94115 USA;Smith-Kettlewell Eye Research Institute, San Francisco, CA 94115 USA

  • Venue:
  • ICPR '06 Proceedings of the 18th International Conference on Pattern Recognition - Volume 04
  • Year:
  • 2006

Quantified Score

Hi-index 0.02

Visualization

Abstract

Much past research on finding text in natural scenes uses bottom-up grouping processes to detect candidate text features as a first processing step. While such grouping procedures are a fast and efficient way of extracting the parts of an image that are most likely to contain text, they still suffer from large amounts of false positives that must be pruned out before they can be read by OCR. We argue that a natural framework for pruning out false positive text features is figure-ground segmentation. This process is implemented using a graphical model (i.e. MRF) in which each candidate text feature is represented by a node. Since each node has only two possible states (figure and ground), and since the connectivity of the graphical model is sparse, we can perform rapid inference on the graph using belief propagation. We show promising results on a variety of urban and indoor scene images containing signs, demonstrating the feasibility of the approach.