Sparse deep belief net for handwritten digits classification

  • Authors:
  • Jiongyun Xie;Hongtao Lu;Deng Nan;Cai Nengbin

  • Affiliations:
  • MOE-MS Key Laboratory for Intelligent Computing and Intelligent Systems, Dept. of Computer Science, Shanghai Jiaotong University, Shanghai, China;MOE-MS Key Laboratory for Intelligent Computing and Intelligent Systems, Dept. of Computer Science, Shanghai Jiaotong University, Shanghai, China;Shanghai Forensic Center;Shanghai Forensic Center

  • Venue:
  • AICI'10 Proceedings of the 2010 international conference on Artificial intelligence and computational intelligence: Part I
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

It has been shown that the Deep Belief Network is good at modeling input distribution, and can be trained efficiently by the greedy layer-wise unsupervised learning. Hoglak Lee et al. (2008) introduced a sparse variant of the Deep Belief Network, which applied the Gaussian linear units to model the input data with a sparsity constraint. However, it takes much more weight updates to train the RBM (Restricted Boltzmann Machine) with Gaussian visible units, and the reconstruction error is much larger than training an RBM with binary visible units. Here, we propose another version of Sparse Deep Belief Net which applies the differentiable sparse coding method to train the first level of the deep network, and then train the higher layers with RBM. This hybrid model, combining the advantage of the Deep architecture and the sparse coding model, leads to state-of-the-art performance on the classification of handwritten digits.