Sciweavers

141 search results - page 8 / 29
» Learning r-of-k Functions by Boosting
Sort
View
EUROCOLT
1995
Springer
15 years 3 months ago
A decision-theoretic generalization of on-line learning and an application to boosting
k. The model we study can be interpreted as a broad, abstract extension of the well-studied on-line prediction model to a general decision-theoretic setting. We show that the multi...
Yoav Freund, Robert E. Schapire
COLT
2008
Springer
15 years 1 months ago
On the Equivalence of Weak Learnability and Linear Separability: New Relaxations and Efficient Boosting Algorithms
Boosting algorithms build highly accurate prediction mechanisms from a collection of lowaccuracy predictors. To do so, they employ the notion of weak-learnability. The starting po...
Shai Shalev-Shwartz, Yoram Singer
ICML
2008
IEEE
16 years 17 days ago
Random classification noise defeats all convex potential boosters
A broad class of boosting algorithms can be interpreted as performing coordinate-wise gradient descent to minimize some potential function of the margins of a data set. This class...
Philip M. Long, Rocco A. Servedio
ICML
2007
IEEE
16 years 17 days ago
On learning with dissimilarity functions
We study the problem of learning a classification task in which only a dissimilarity function of the objects is accessible. That is, data are not represented by feature vectors bu...
Liwei Wang, Cheng Yang, Jufu Feng
ICML
2004
IEEE
16 years 17 days ago
Surrogate maximization/minimization algorithms for AdaBoost and the logistic regression model
Surrogate maximization (or minimization) (SM) algorithms are a family of algorithms that can be regarded as a generalization of expectation-maximization (EM) algorithms. There are...
Zhihua Zhang, James T. Kwok, Dit-Yan Yeung