Links between probabilistic automata and hidden Markov models: probability distributions, learning models and induction algorithms

  • Authors:
  • P. Dupont;F. Denis;Y. Esposito

  • Affiliations:
  • INGI, Université catholique de Louvain, Place Sainte-Barbe 2, B-1348 Louvain-la-Neuve, Belgium;LIF-CMI, UMR 6166, 39, Rue F. Joliot Curie, 13453 Marseille Cedex 13, France;LIF-CMI, UMR 6166, 39, Rue F. Joliot Curie, 13453 Marseille Cedex 13, France

  • Venue:
  • Pattern Recognition
  • Year:
  • 2005

Quantified Score

Hi-index 0.01

Visualization

Abstract

This article presents an overview of Probabilistic Automata (PA) and discrete Hidden Markov Models (HMMs), and aims at clarifying the links between them. The first part of this work concentrates on probability distributions generated by these models. Necessary and sufficient conditions for an automaton to define a probabilistic language are detailed. It is proved that probabilistic deterministic automata (PDFA) form a proper subclass of probabilistic non-deterministic automata (PNFA). Two families of equivalent models are described next. On one hand, HMMs and PNFA with no final probabilities generate distributions over complete finite prefix-free sets. On the other hand, HMMs with final probabilities and probabilistic automata generate distributions over strings of finite length. The second part of this article presents several learning models, which formalize the problem of PA induction or, equivalently, the problem of HMM topology induction and parameter estimation. These learning models include the PAC and identification with probability 1 frameworks. Links with Bayesian learning are also discussed. The last part of this article presents an overview of induction algorithms for PA or HMMs using state merging, state splitting, parameter pruning and error-correcting techniques.