Basic principles of learning Bayesian logic programs

  • Authors:
  • Kristian Kersting;Luc De Raedt

  • Affiliations:
  • CSAIL, Massachusetts Institute of Technologie, Cambridge, MA;Departement Computerwetenschappen, K.U. Leuven, Heverlee, Belgium

  • Venue:
  • Probabilistic inductive logic programming
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

Bayesian logic programs tightly integrate definite logic programs with Bayesian networks in order to incorporate the notions of objects and relations into Bayesian networks. They establish a one-to-one mapping between ground atoms and random variables, and between the immediate consequence operator and the directly influenced by relation. In doing so, they nicely separate the qualitative (i.e. logical) component from the quantitative (i.e. the probabilistic) one providing a natural framework to describe general, probabilistic dependencies among sets of random variables. In this chapter, we present results on combining Inductive Logic Programming with Bayesian networks to learn both the qualitative and the quantitative components of Bayesian logic programs from data. More precisely, we show how the qualitative components can be learned by combining the inductive logic programming setting learning from interpretations with score-based techniques for learning Bayesian networks. The estimation of the quantitative components is reduced to the corresponding problem of (dynamic) Bayesian networks.