Robust alternating AdaBoost

  • Authors:
  • Héctor Allende-Cid;Rodrigo Salas;Héctor Allende;Ricardo Ñanculef

  • Affiliations:
  • Universidad Técnica Federico Santa María, Dept. de Informática, Casilla, Valparaíso, Chile;Universidad de Valparaíso, Departamento de Ingeniería Biomédica, Valparaíso, Chile;Universidad Técnica Federico Santa María, Dept. de Informática, Valparaíso, Chile;Universidad Técnica Federico Santa María, Dept. de Informática, Valparaíso, Chile

  • Venue:
  • CIARP'07 Proceedings of the Congress on pattern recognition 12th Iberoamerican conference on Progress in pattern recognition, image analysis and applications
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

Ensemble methods are general techniques to improve the accuracy of any given learning algorithm. Boosting is a learning algorithm that builds the classifier ensembles incrementally. In this work we propose an improvement of the classical and inverse AdaBoost algorithms to deal with the problem of the presence of outliers in the data. We propose the Robust Alternating AdaBoost (RADA) algorithm that alternates between the classic and inverse AdaBoost to create a more stable algorithm. The RADA algorithm bounds the influence of the outliers to the empirical distribution, it detects and diminishes the empirical probability of "bad" samples, and it performs a more accurate classification under contaminated data. We report the performance results using synthetic and real datasets, the latter obtained from a benchmark site.