Detecting visual text

  • Authors:
  • Jesse Dodge;Amit Goyal;Xufeng Han;Alyssa Mensch;Margaret Mitchell;Karl Stratos;Kota Yamaguchi;Yejin Choi;Hal Daumé, III;Alexander C. Berg;Tamara L. Berg

  • Affiliations:
  • University of Washington;University of Maryland;Stony Brook University;MIT;Oregon Health & Science University;Columbia University;Stony Brook University;Stony Brook University;University of Maryland;Stony Brook University;Stony Brook University

  • Venue:
  • NAACL HLT '12 Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

When people describe a scene, they often include information that is not visually apparent; sometimes based on background knowledge, sometimes to tell a story. We aim to separate visual text---descriptions of what is being seen---from non-visual text in natural images and their descriptions. To do so, we first concretely define what it means to be visual, annotate visual text and then develop algorithms to automatically classify noun phrases as visual or non-visual. We find that using text alone, we are able to achieve high accuracies at this task, and that incorporating features derived from computer vision algorithms improves performance. Finally, we show that we can reliably mine visual nouns and adjectives from large corpora and that we can use these effectively in the classification task.