Sciweavers

Share
NIPS
1998

Lazy Learning Meets the Recursive Least Squares Algorithm

9 years 11 months ago
Lazy Learning Meets the Recursive Least Squares Algorithm
Lazy learning is a memory-based technique that, once a query is received, extracts a prediction interpolating locally the neighboring examples of the query which are considered relevant according to a distance measure. In this paper we propose a datadriven method to select on a query-by-query basis the optimal number of neighbors to be considered for each prediction. As an efficient way to identify and validate local models, the recursive least squares algorithm is introduced in the context of local approximation and lazy learning. Furthermore, beside the winner-takes-all strategy for model selection, a local combination of the most promising models is explored. The method proposed is tested on six different datasets and compared with a state-of-the-art approach.
Mauro Birattari, Gianluca Bontempi, Hugues Bersini
Added 01 Nov 2010
Updated 01 Nov 2010
Type Conference
Year 1998
Where NIPS
Authors Mauro Birattari, Gianluca Bontempi, Hugues Bersini
Comments (0)
books