Mixed Memory Markov Models: Decomposing Complex Stochastic Processes as Mixtures of Simpler Ones

  • Authors:
  • Lawrence K. Saul;Michael I. Jordan

  • Affiliations:
  • AT&T Labs, Florham Park, NJ 07932. lsaul@research.att.com;University of California, Berkeley, CA 94720. jordan@cs.berkeley.edu

  • Venue:
  • Machine Learning
  • Year:
  • 1999

Quantified Score

Hi-index 0.00

Visualization

Abstract

We study Markov models whose state spaces arise from the Cartesian product of two or more discrete random variables. We show how to parameterize the transition matrices of these models as a convex combination—or mixture—of simpler dynamical models. The parameters in these models admit a simple probabilistic interpretation and can be fitted iteratively by an Expectation-Maximization (EM) procedure. We derive a set of generalized Baum-Welch updates for factorial hidden Markov models that make use of this parameterization. We also describe a simple iterative procedure for approximately computing the statistics of the hidden states. Throughout, we give examples where mixed memory models provide a useful representation of complex stochastic processes.