Complexity of Comparing Hidden Markov Models

  • Authors:
  • Rune B. Lyngsø;Christian N. S. Pedersen

  • Affiliations:
  • -;-

  • Venue:
  • ISAAC '01 Proceedings of the 12th International Symposium on Algorithms and Computation
  • Year:
  • 2001

Quantified Score

Hi-index 0.02

Visualization

Abstract

The basic theory of hidden Markov models was developed and applied to problems in speech recognition in the late 1960's, and has since then been applied to numerous problems, e.g. biological sequence analysis. In this paper we consider the problem of computing the most likely string generated by a given model, and its implications on the complexity of comparing hidden Markov models. We show that computing the most likely string, and approximating its probability within any constant factor, is NP-hard, and establish the NP-hardness of comparing two hidden Markov models under the L∞- and L1-norms. We discuss the applicability of the technique used to other measures of distance between probability distributions. In particular we show that it cannot be used to prove NP-hardness of determining the Kullback-Leibler distance between the probability distributions of two hidden Markov models, or of comparing them under the Lk-norm for any fixed even integer k.