A bounded version of online boosting on open-ended data streams

  • Authors:
  • José Luis Triviño-Rodriguez;Amparo Ruiz-Sepúlveda;Rafael Morales-Bueno

  • Affiliations:
  • Department of Computer Science and Artificial Intelligence, University of Málaga, Málaga, Spain;Department of Computer Science and Artificial Intelligence, University of Málaga, Málaga, Spain;Department of Computer Science and Artificial Intelligence, University of Málaga, Málaga, Spain

  • Venue:
  • DaWaK'11 Proceedings of the 13th international conference on Data warehousing and knowledge discovery
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we propose an incremental learning algorithm for ensemble classifier systems. Ensemble learning algorithms combine the predictions of multiple base models, each of which is learned using a traditional algorithm. We propose a new method to update weights of classifiers in the weighted majority voting scheme under the one-pass incremental learning situations. This method computes the weights of classifiers and the distribution of training data following an approach based on the computing of prequential error that avoids the overflow of internal values used by the learning algorithm. Using a prequential approach implies that learned samples are forgotten progressively. Forgetting learned concepts could influence the accuracy of the model. However, in the experiments, we verify that the proposed model can learn incrementally without serious forgetting and that the performance is not seriously influenced by the used reweighting method in comparison with learning models without forgetting. Experimental results confirm that the proposed incremental ensemble classifier system yields comparable performance with another learning ensemble classifier system. Moreover, it can be trained with open-ended data streams without data overflow.