Metadata for mixed-media access

  • Authors:
  • Francine Chen;Marti Hearst;Julian Kupiec;Jan Pedersen;Lynn Wilcox

  • Affiliations:
  • Xerox Palo Alto Research Center, 3333 Coyote Hill Road, Palo Alto, CA;Xerox Palo Alto Research Center, 3333 Coyote Hill Road, Palo Alto, CA;Xerox Palo Alto Research Center, 3333 Coyote Hill Road, Palo Alto, CA;Xerox Palo Alto Research Center, 3333 Coyote Hill Road, Palo Alto, CA;Xerox Palo Alto Research Center, 3333 Coyote Hill Road, Palo Alto, CA

  • Venue:
  • ACM SIGMOD Record
  • Year:
  • 1994

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we discuss mixed-media access, an information access paradigm for multimedia data in which the media type of a query may differ from that of the data. The types of media considered in this paper are speech, images of text, and full-length text. Some examples of metadata for mixed-media access are locations of keywords in speech and images, identification of speakers, locations of emphasized regions in speech, and locations of topic boundaries in text. Algorithms for automatically generating this metadata are described, including word spotting, speaker segmentation, emphatic speech detection, and subtopic boundary location. We illustrate queries composed of diverse media types in an example of access to recorded meetings, via speaker and keyword location.