The kernel recursive least-squares algorithm

  • Authors:
  • Y. Engel;S. Mannor;R. Meir

  • Affiliations:
  • Center for Neural Comput., Hebrew Univ., Jerusalem, Israel;-;-

  • Venue:
  • IEEE Transactions on Signal Processing
  • Year:
  • 2004

Quantified Score

Hi-index 35.70

Visualization

Abstract

We present a nonlinear version of the recursive least squares (RLS) algorithm. Our algorithm performs linear regression in a high-dimensional feature space induced by a Mercer kernel and can therefore be used to recursively construct minimum mean-squared-error solutions to nonlinear least-squares problems that are frequently encountered in signal processing applications. In order to regularize solutions and keep the complexity of the algorithm bounded, we use a sequential sparsification process that admits into the kernel representation a new input sample only if its feature space image cannot be sufficiently well approximated by combining the images of previously admitted samples. This sparsification procedure allows the algorithm to operate online, often in real time. We analyze the behavior of the algorithm, compare its scaling properties to those of support vector machines, and demonstrate its utility in solving two signal processing problems-time-series prediction and channel equalization.