Deriving image-text document surrogates to optimize cognition

  • Authors:
  • Eunyee Koh;Andruid Kerne

  • Affiliations:
  • Adobe Systems Inc, San Jose, CA, USA;Texas A&M University, College Station, TX, USA

  • Venue:
  • Proceedings of the 9th ACM symposium on Document engineering
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

The representation of information collections needs to be optimized for human cognition. While documents often include rich visual components, collections, including personal collections and those generated by search engines, are typically represented by lists of text-only surrogates. By concurrently invoking complementary components of human cognition, combined image-text surrogates will help people to more effectively see, understand, think about, and remember an information collection. This research develops algorithmic methods that use the structural context of images in HTML documents to associate meaningful text and thus derive combined image-text surrogates. Our algorithm first recognizes which documents consist essentially of informative and multimedia content. Then, the algorithm recognizes the informative sub-trees within each such document, discards advertisements and navigation, and extracts images with contextual descriptions. Experimental results demonstrate the algorithm's efficacy. An implementation of the algorithm is provided in combinFormation, a creativity support tool for collection authoring. The enhanced image-text surrogates enhance the experiences of users finding and collecting information as part of developing new ideas.