Recognizing context for annotating a live life recording

  • Authors:
  • Nicky Kern;Bernt Schiele;Albrecht Schmidt

  • Affiliations:
  • Department of Computer Science, Darmstadt University of Technology, Darmstadt, Germany;Department of Computer Science, Darmstadt University of Technology, Darmstadt, Germany;Embedded Interaction Research Group, Universität München, München, Germany

  • Venue:
  • Personal and Ubiquitous Computing - Memory and Sharing of Experiences
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

In the near future, it will be possible to continuously record and store the entire audio---visual lifetime of a person together with all digital information that the person perceives or creates. While the storage of this data will be possible soon, retrieval and indexing into such large data sets are unsolved challenges. Since today's retrieval cues seem insufficient we argue that additional cues, obtained from body-worn sensors, make associative retrieval by humans possible. We present three approaches to create such cues, each along with an experimental evaluation: the user's physical activity from acceleration sensors, his social environment from audio sensors, and his interruptibility from multiple sensors.