Effect of pruning and early stopping on performance of a boosting ensemble

  • Authors:
  • Harris Drucker

  • Affiliations:
  • Monmouth University, West Long Branch, NJ

  • Venue:
  • Computational Statistics & Data Analysis - Nonlinear methods and data mining
  • Year:
  • 2002

Quantified Score

Hi-index 0.01

Visualization

Abstract

Generating an architecture for an ensemble of boosting machines involves making a series of design decisions. One design decision is whether to use simple "weak learners" such as decision tree stumps or more complicated weak learners such as large decision trees or neural networks. Another design decision is the training algorithm for the constituent weak learners. Here we concentrate on binary decision trees and show that the best results are obtained using the Z-criterion to build the trees without pruning. In using neural networks, early stopping is recommended as an approach to lower the training time. In examining the multi-class boosting algorithms, the jury is still out on whether using the all-pairs binary learning algorithm or pseudo-loss is better.