We consider the AdaBoost procedure for boosting weak learners. In AdaBoost, a key step is choosing a new distribution on the training examples based on the old distribution and th...
We present a novel boosting algorithm, called SoftBoost, designed for sets of binary labeled examples that are not necessarily separable by convex combinations of base hypotheses....
Manfred K. Warmuth, Karen A. Glocer, Gunnar Rä...
Recently, boosting has come to be used widely in object-detection applications because of its impressive performance in both speed and accuracy. However, learning weak classifier...
Recently, boosting is used widely in object detection applications because of its impressive performance in both speed and accuracy. However, learning weak classifiers which is on...
We consider boosting algorithms that maintain a distribution over a set of examples. At each iteration a weak hypothesis is received and the distribution is updated. We motivate t...