Sciweavers

13 search results - page 1 / 3
» The Importance of Convexity in Learning with Squared Loss
Sort
View
TIT
1998
70views more  TIT 1998»
13 years 4 months ago
The Importance of Convexity in Learning with Squared Loss
We show that if the closureof a function class F under the metric induced by some probability distribution is not convex, then the sample complexity for agnostically learning F wi...
Wee Sun Lee, Peter L. Bartlett, Robert C. Williams...
NIPS
2008
13 years 6 months ago
On the Efficient Minimization of Classification Calibrated Surrogates
Bartlett et al (2006) recently proved that a ground condition for convex surrogates, classification calibration, ties up the minimization of the surrogates and classification risk...
Richard Nock, Frank Nielsen
SDM
2008
SIAM
150views Data Mining» more  SDM 2008»
13 years 6 months ago
A Stagewise Least Square Loss Function for Classification
This paper presents a stagewise least square (SLS) loss function for classification. It uses a least square form within each stage to approximate a bounded monotonic nonconvex los...
Shuang-Hong Yang, Bao-Gang Hu
ACL
2008
13 years 6 months ago
Semi-Supervised Convex Training for Dependency Parsing
We present a novel semi-supervised training algorithm for learning dependency parsers. By combining a supervised large margin loss with an unsupervised least squares loss, a discr...
Qin Iris Wang, Dale Schuurmans, Dekang Lin
JMLR
2012
11 years 7 months ago
Beyond Logarithmic Bounds in Online Learning
We prove logarithmic regret bounds that depend on the loss L∗ T of the competitor rather than on the number T of time steps. In the general online convex optimization setting, o...
Francesco Orabona, Nicolò Cesa-Bianchi, Cla...