The value of agreement a new boosting algorithm

  • Authors:
  • Boaz Leskes;Leen Torenvliet

  • Affiliations:
  • University of Amsterdam, Department of Computer Science, Plantage Muidergracht 24, 1018 TV Amsterdam, Netherlands;University of Amsterdam, Department of Computer Science, Plantage Muidergracht 24, 1018 TV Amsterdam, Netherlands

  • Venue:
  • Journal of Computer and System Sciences
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

In the past few years unlabeled examples and their potential advantage have received a lot of attention. In this paper a new boosting algorithm is presented where unlabeled examples are used to enforce agreement between several different learning algorithms. Not only do the learning algorithms learn from the given training set but they are supposed to do so while agreeing on the unlabeled examples. Similar ideas have been proposed before (for example, the Co-Training algorithm by Mitchell and Blum), but without a proof or under strong assumptions. In our setting, it is only assumed that all learning algorithms are equally adequate for the tasks. A new generalization bound is presented where the use of unlabeled examples results in a better ratio between training-set size and the resulting classifier's quality and thus reduce the number of labeled examples necessary for achieving it. The extent of this improvement depends on the diversity of the learners-a more diverse group of learners will result in a larger improvement whereas using two copies of a single algorithm gives no advantage at all. As a proof of concept, the algorithm, named Agreement Boost, is applied to two test problems. In both cases, using Agreement Boost results in an up to 40% reduction in the number of labeled examples.