Sciweavers

65 search results - page 2 / 13
» Gradient Descent for General Reinforcement Learning
Sort
View
GECCO
2006
Springer
159views Optimization» more  GECCO 2006»
13 years 9 months ago
Standard and averaging reinforcement learning in XCS
This paper investigates reinforcement learning (RL) in XCS. First, it formally shows that XCS implements a method of generalized RL based on linear approximators, in which the usu...
Pier Luca Lanzi, Daniele Loiacono
ICML
2009
IEEE
14 years 6 months ago
Gradient descent with sparsification: an iterative algorithm for sparse recovery with restricted isometry property
We present an algorithm for finding an ssparse vector x that minimizes the squareerror y - x 2 where satisfies the restricted isometry property (RIP), with isometric constant 2s ...
Rahul Garg, Rohit Khandekar
JMLR
2012
11 years 7 months ago
Krylov Subspace Descent for Deep Learning
In this paper, we propose a second order optimization method to learn models where both the dimensionality of the parameter space and the number of training samples is high. In ou...
Oriol Vinyals, Daniel Povey
NIPS
2007
13 years 6 months ago
Incremental Natural Actor-Critic Algorithms
We present four new reinforcement learning algorithms based on actor-critic and natural-gradient ideas, and provide their convergence proofs. Actor-critic reinforcement learning m...
Shalabh Bhatnagar, Richard S. Sutton, Mohammad Gha...
ICDM
2010
IEEE
167views Data Mining» more  ICDM 2010»
13 years 3 months ago
Averaged Stochastic Gradient Descent with Feedback: An Accurate, Robust, and Fast Training Method
On large datasets, the popular training approach has been stochastic gradient descent (SGD). This paper proposes a modification of SGD, called averaged SGD with feedback (ASF), tha...
Xu Sun, Hisashi Kashima, Takuya Matsuzaki, Naonori...