Learning Kernel Perceptrons on Noisy Data Using Random Projections

  • Authors:
  • Guillaume Stempfel;Liva Ralaivola

  • Affiliations:
  • Laboratoire d'Informatique Fondamentale de Marseille, Umr Cnrs 6166, Université de Provence, 39, rue Joliot Curie, 13013 Marseille, France;Laboratoire d'Informatique Fondamentale de Marseille, Umr Cnrs 6166, Université de Provence, 39, rue Joliot Curie, 13013 Marseille, France

  • Venue:
  • ALT '07 Proceedings of the 18th international conference on Algorithmic Learning Theory
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we address the issue of learning nonlinearly separable concepts with a kernel classifier in the situation where the data at hand are altered by a uniform classification noise. Our proposed approach relies on the combination of the technique of random or deterministic projections with a classification noise tolerant perceptron learning algorithm that assumes distributions defined over finite-dimensional spaces. Provided a sufficient separation margin characterizes the problem, this strategy makes it possible to envision the learning from a noisy distribution in any separable Hilbert space, regardless of its dimension; learning with any appropriate Mercer kernel is therefore possible. We prove that the required sample complexity and running time of our algorithm is polynomial in the classical PAC learning parameters. Numerical simulations on toy datasets and on data from the UCI repository support the validity of our approach.