Training restricted boltzmann machines with multi-tempering: harnessing parallelization

  • Authors:
  • Philemon Brakel;Sander Dieleman;Benjamin Schrauwen

  • Affiliations:
  • Department of Electronics and Information Systems, Ghent University, Gent, Belgium;Department of Electronics and Information Systems, Ghent University, Gent, Belgium;Department of Electronics and Information Systems, Ghent University, Gent, Belgium

  • Venue:
  • ICANN'12 Proceedings of the 22nd international conference on Artificial Neural Networks and Machine Learning - Volume Part II
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

Restricted Boltzmann Machines (RBM's) are unsupervised probabilistic neural networks that can be stacked to form Deep Belief Networks. Given the recent popularity of RBM's and the increasing availability of parallel computing architectures, it becomes interesting to investigate learning algorithms for RBM's that benefit from parallel computations. In this paper, we look at two extensions of the parallel tempering algorithm, which is a Markov Chain Monte Carlo method to approximate the likelihood gradient. The first extension is directed at a more effective exchange of information among the parallel sampling chains. The second extension estimates gradients by averaging over chains from different temperatures. We investigate the efficiency of the proposed methods and demonstrate their usefulness on the MNIST dataset. Especially the weighted averaging seems to benefit Maximum Likelihood learning.