Distributed learning with bagging-like performance

  • Authors:
  • Nitesh V. Chawla;Thomas E. Moore;Lawrence O. Hall;Kevin W. Bowyer;W. Philip Kegelmeyer;Clayton Springer

  • Affiliations:
  • Department of Computer Science and Engineering, University of South Florida, 4202 East Flower Avenue, Tampa, FL;Department of Computer Science and Engineering, University of South Florida, 4202 East Flower Avenue, Tampa, FL;Department of Computer Science and Engineering, University of South Florida, 4202 East Flower Avenue, Tampa, FL;Department of Computer Science and Engineering, University of Notre Dame, Notre Dame, IN;Sandia National Laboratories, Biosystems Research Department, P.O. Box 969, MS 9951, Livermore, CA;Sandia National Laboratories, Biosystems Research Department, P.O. Box 969, MS 9951, Livermore, CA

  • Venue:
  • Pattern Recognition Letters
  • Year:
  • 2003

Quantified Score

Hi-index 0.10

Visualization

Abstract

Bagging forms a committee of classifiers by bootstrap aggregation of training sets from a pool of training data. A simple alternative to bagging is to partition the data into disjoint subsets. Experiments with decision tree and neural network classifiers on various datasets show that, given the same size partitions and bags, disjoint partitions result in performance equivalent to, or better than, bootstrap aggregates (bags). Many applications (e.g., protein structure prediction) involve use of datasets that are too large to handle in the memory of the typical computer. Hence, bagging with samples the size of the data is impractical. Our results indicate that, in such applications, the simple approach of creating a committee of n classifiers from disjoint partitions each of size 1/n (which will be memory resident during learning) in a distributed way results in a classifier which has a bagging-like performance gain. The use of distributed disjoint partitions in learning is significantly less complex and faster than bagging.