Sciweavers

PR
2008

Quadratic boosting

13 years 4 months ago
Quadratic boosting
This paper presents a strategy to improve the AdaBoost algorithm with a quadratic combination of base classifiers. We observe that learning this combination is necessary to get better performance and is possible by constructing an intermediate learner operating on the combined linear and quadratic terms. This is not trivial, as the parameters of the base classifiers are not under direct control, obstructing the application of direct optimization. We propose a new method realizing iterative optimization indirectly. First we train a classifier by randomizing the labels of training examples. Subsequently, the input learner is called repeatedly with a systematic update of the labels of the training examples in each round. We show that the quadratic boosting algorithm converges under the condition that the given base learner minimizes the empirical error. We also give an upper bound on the VC-dimension of the new classifier. Our experimental results on 23 standard problems show that quadra...
Thang V. Pham, Arnold W. M. Smeulders
Added 14 Dec 2010
Updated 14 Dec 2010
Type Journal
Year 2008
Where PR
Authors Thang V. Pham, Arnold W. M. Smeulders
Comments (0)