Sciweavers

146 search results - page 1 / 30
» Online Gradient Descent Learning Algorithms
Sort
View
FOCM
2008
140views more  FOCM 2008»
13 years 4 months ago
Online Gradient Descent Learning Algorithms
This paper considers the least-square online gradient descent algorithm in a reproducing kernel Hilbert space (RKHS) without explicit regularization. We present a novel capacity i...
Yiming Ying, Massimiliano Pontil
CORR
2004
Springer
103views Education» more  CORR 2004»
13 years 4 months ago
Online convex optimization in the bandit setting: gradient descent without a gradient
We study a general online convex optimization problem. We have a convex set S and an unknown sequence of cost functions c1, c2, . . . , and in each period, we choose a feasible po...
Abraham Flaxman, Adam Tauman Kalai, H. Brendan McM...
COLT
2010
Springer
13 years 2 months ago
Composite Objective Mirror Descent
We present a new method for regularized convex optimization and analyze it under both online and stochastic optimization settings. In addition to unifying previously known firstor...
John Duchi, Shai Shalev-Shwartz, Yoram Singer, Amb...
JMLR
2006
116views more  JMLR 2006»
13 years 4 months ago
Step Size Adaptation in Reproducing Kernel Hilbert Space
This paper presents an online support vector machine (SVM) that uses the stochastic meta-descent (SMD) algorithm to adapt its step size automatically. We formulate the online lear...
S. V. N. Vishwanathan, Nicol N. Schraudolph, Alex ...
ACL
2009
13 years 2 months ago
Stochastic Gradient Descent Training for L1-regularized Log-linear Models with Cumulative Penalty
Stochastic gradient descent (SGD) uses approximate gradients estimated from subsets of the training data and updates the parameters in an online fashion. This learning framework i...
Yoshimasa Tsuruoka, Jun-ichi Tsujii, Sophia Anania...