Publication | Closed Access
The Kernel Recursive Least-Squares Algorithm
1.1K
Citations
35
References
2004
Year
Mercer KernelSupport Vector MachineEngineeringMachine LearningData ScienceRecursive Least SquaresPattern RecognitionSparse RepresentationReproducing Kernel MethodInverse ProblemsComputer ScienceStatistical Learning TheoryNonlinear VersionApproximation TheorySignal ProcessingKernel Method
We present a nonlinear version of the recursive least squares (RLS) algorithm. Our algorithm performs linear regression in a high-dimensional feature space induced by a Mercer kernel and can therefore be used to recursively construct minimum mean-squared-error solutions to nonlinear least-squares problems that are frequently encountered in signal processing applications. In order to regularize solutions and keep the complexity of the algorithm bounded, we use a sequential sparsification process that admits into the kernel representation a new input sample only if its feature space image cannot be sufficiently well approximated by combining the images of previously admitted samples. This sparsification procedure allows the algorithm to operate online, often in real time. We analyze the behavior of the algorithm, compare its scaling properties to those of support vector machines, and demonstrate its utility in solving two signal processing problems-time-series prediction and channel equalization.
| Year | Citations | |
|---|---|---|
Page 1
Page 1