Dynamics of variance reduction in bagging and other techniques based on randomisation

  • Authors:
  • G. Fumera;F. Roli;A. Serrau

  • Affiliations:
  • Dept. of Electrical and Electronic Eng., University of Cagliari, Italy;Dept. of Electrical and Electronic Eng., University of Cagliari, Italy;Dept. of Electrical and Electronic Eng., University of Cagliari, Italy

  • Venue:
  • MCS'05 Proceedings of the 6th international conference on Multiple Classifier Systems
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper the performance of bagging in classification problems is theoretically analysed, using a framework developed in works by Tumer and Ghosh and extended by the authors. A bias-variance decomposition is derived, which relates the expected misclassification probability attained by linearly combining classifiers trained on N bootstrap replicates of a fixed training set to that attained by a single bootstrap replicate of the same training set. Theoretical results show that the expected misclassification probability of bagging has the same bias component as a single bootstrap replicate, while the variance component is reduced by a factor N. Experimental results show that the performance of bagging as a function of the number of bootstrap replicates follows quite well our theoretical prediction. It is finally shown that theoretical results derived for bagging also apply to other methods for constructing multiple classifiers based on randomisation, such as the random subspace method and tree randomisation.