Robustness of Kernel Based Regression: A Comparison of Iterative Weighting Schemes

  • Authors:
  • Kris Brabanter;Kristiaan Pelckmans;Jos Brabanter;Michiel Debruyne;Johan A. Suykens;Mia Hubert;Bart Moor

  • Affiliations:
  • KULeuven, ESAT-SCD, Leuven, Belgium 3001;KULeuven, ESAT-SCD, Leuven, Belgium 3001;KULeuven, ESAT-SCD, Leuven, Belgium 3001 and Departement Ind. Ing., KaHo Sint-Lieven (Associatie K.U.Leuven), Gent B-9000;Department of Mathematics and Computer Science, Universiteit Antwerpen, Antwerpen, Belgium B-2020;KULeuven, ESAT-SCD, Leuven, Belgium 3001;Department of Statistics, KULeuven, Leuven, Belgium B-3001;KULeuven, ESAT-SCD, Leuven, Belgium 3001

  • Venue:
  • ICANN '09 Proceedings of the 19th International Conference on Artificial Neural Networks: Part I
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

It has been shown that Kernel Based Regression (KBR) with a least squares loss has some undesirable properties from robustness point of view. KBR with more robust loss functions, e.g. Huber or logistic losses, often give rise to more complicated computations. In this work the practical consequences of this sensitivity are explained, including the breakdown of Support Vector Machines (SVM) and weighted Least Squares Support Vector Machines (LS-SVM) for regression. In classical statistics, robustness is improved by reweighting the original estimate. We study the influence of reweighting the LS-SVM estimate using four different weight functions. Our results give practical guidelines in order to choose the weights, providing robustness and fast convergence. It turns out that Logistic and Myriad weights are suitable reweighting schemes when outliers are present in the data. In fact, the Myriad shows better performance over the others in the presence of extreme outliers (e.g. Cauchy distributed errors). These findings are then illustrated on toy example as well as on a real life data sets.