Generalization Performance of Subspace Bayes Approach in Linear Neural Networks

  • Authors:
  • Shinichi Nakajima;Sumio Watanabe

  • Affiliations:
  • The authors are with Tokyo Institute of Technology, Yokohama-shi, 226--8503 Japan. E-mail: nakajima.s@cs.pi.titech.ac.jp, E-mail: swatanab@pi.titech.ac.jp,;The authors are with Tokyo Institute of Technology, Yokohama-shi, 226--8503 Japan. E-mail: nakajima.s@cs.pi.titech.ac.jp, E-mail: swatanab@pi.titech.ac.jp,

  • Venue:
  • IEICE - Transactions on Information and Systems
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

In unidentifiable models, the Bayes estimation has the advantage of generalization performance over the maximum likelihood estimation. However, accurate approximation of the posterior distribution requires huge computational costs. In this paper, we consider an alternative approximation method, which we call a subspace Bayes approach. A subspace Bayes approach is an empirical Bayes approach where a part of the parameters are regarded as hyperparameters. Consequently, in some three-layer models, this approach requires much less computational costs than Markov chain Monte Carlo methods. We show that, in three-layer linear neural networks, a subspace Bayes approach is asymptotically equivalent to a positive-part James-Stein type shrinkage estimation, and theoretically clarify its generalization error and training error. We also discuss the domination over the maximum likelihood estimation and the relation to the variational Bayes approach.