Bayesian Model Learning Based on Predictive Entropy

  • Authors:
  • Jukka Corander;Pekka Marttinen

  • Affiliations:
  • Department of Mathematics and Statistics, University of Helsinki, Helsinki, Finland FIN-00014;Department of Mathematics and Statistics, University of Helsinki, Helsinki, Finland FIN-00014

  • Venue:
  • Journal of Logic, Language and Information
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

Bayesian paradigm has been widely acknowledged as a coherent approach to learning putative probability model structures from a finite class of candidate models. Bayesian learning is based on measuring the predictive ability of a model in terms of the corresponding marginal data distribution, which equals the expectation of the likelihood with respect to a prior distribution for model parameters. The main controversy related to this learning method stems from the necessity of specifying proper prior distributions for all unknown parameters of a model, which ensures a complete determination of the marginal data distribution. Even for commonly used models, subjective priors may be difficult to specify precisely, and therefore, several automated learning procedures have been suggested in the literature. Here we introduce a novel Bayesian learning method based on the predictive entropy of a probability model, that can combine both subjective and objective probabilistic assessment of uncertain quantities in putative models. It is shown that our approach can avoid some of the limitations of the earlier suggested objective Bayesian methods.