Parallel Learning of Belief Networks in Large and Difficult Domains

  • Authors:
  • Y. Xiang;T. Chu

  • Affiliations:
  • Department of Computer Science, University of Regina, Regina, Saskatchewan, Canada S4S 0A2. yxiang@cs.uregina.ca;Avant Corporation, Sunnyvale, CA, USA

  • Venue:
  • Data Mining and Knowledge Discovery
  • Year:
  • 1999

Quantified Score

Hi-index 0.01

Visualization

Abstract

Learning belief networks from large domains can beexpensive even with single-link lookahead search (SLLS). Since aSLLS cannot learn correctly in a class of problem domains, multi-linklookahead search (MLLS) is needed which further increases thecomputational complexity. In our experiment, learning in somedifficult domains over more than a dozen variables took days. In thispaper, we study how to use parallelism to speed up SLLS for learningin large domains and to tackle the increased complexity of MLLS forlearning in difficult domains.We propose a natural decomposition of the learning task for parallelprocessing. We investigate two strategies for job allocation amongprocessors to further improve load balancing and efficiency of theparallel system. For learning from very large datasets, we present aregrouping of the available processors such that slow data accessthrough the file system can be replaced by fast memory access.Experimental results in a distributed memory MIMD computerdemonstrate the effectiveness of the proposed algorithms.