Learning large scale common sense models of everyday life

  • Authors:
  • William Pentney;Matthai Philipose;Jeff Bilmes;Henry Kautz

  • Affiliations:
  • Department of Computer Science & Engineering, University of Washington;Intel Research Seattle;Department of Electrical Engineering, University of Washington;Department of Computer Science, University of Rochester

  • Venue:
  • AAAI'07 Proceedings of the 22nd national conference on Artificial intelligence - Volume 1
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

Recent work has shown promise in using large, publicly available, hand-contributed commonsense databases as joint models that can be used to infer human state from day-to-day sensor data. The parameters of these models are mined from the web. We show in this paper that learning these parameters using sensor data (with the mined parameters as priors) can improve performance of the models significantly. The primary challenge in learning is scale. Since the model comprises roughly 50,000 irregularly connected nodes in each time slice, it is intractable either to completely label observed data manually or to compute the expected likelihood of even a single lime slice. We show how to solve the resulting semi-supervised learning problem by combining a variety of conventional approximation techniques and a novel technique for simplifying the model called context-based pruning. We show empirically that the learned model is substantially better at interpreting sensor data and an detailed analysis of how various techniques contribute to the performance.