Sciweavers

Share
ALT
2007
Springer

Learning Kernel Perceptrons on Noisy Data Using Random Projections

12 years 3 months ago
Learning Kernel Perceptrons on Noisy Data Using Random Projections
In this paper, we address the issue of learning nonlinearly separable concepts with a kernel classifier in the situation where the data at hand are altered by a uniform classification noise. Our proposed approach relies on the combination of the technique of random or deterministic projections with a classification noise tolerant perceptron learning algorithm that assumes distributions defined over finite-dimensional spaces. Provided a sufficient separation margin characterizes the problem, this strategy makes it possible to envision the learning from a noisy distribution in any separable Hilbert space, regardless of its dimension; learning with any appropriate Mercer kernel is therefore possible. We prove that the required sample complexity and running time of our algorithm is polynomial in the classical PAC learning parameters. Numerical simulations on toy datasets and on data from the UCI repository support the validity of our approach.
Guillaume Stempfel, Liva Ralaivola
Added 14 Mar 2010
Updated 14 Mar 2010
Type Conference
Year 2007
Where ALT
Authors Guillaume Stempfel, Liva Ralaivola
Comments (0)
books