Sciweavers

288 search results - page 5 / 58
» Learning to Play Chess Using Temporal Differences
Sort
View
NIPS
2008
13 years 7 months ago
Temporal Difference Based Actor Critic Learning - Convergence and Neural Implementation
Actor-critic algorithms for reinforcement learning are achieving renewed popularity due to their good convergence properties in situations where other approaches often fail (e.g.,...
Dotan Di Castro, Dmitry Volkinshtein, Ron Meir
ICCBR
2010
Springer
13 years 9 months ago
Reducing the Memory Footprint of Temporal Difference Learning over Finitely Many States by Using Case-Based Generalization
In this paper we present an approach for reducing the memory footprint requirement of temporal difference methods in which the set of states is finite. We use case-based generaliza...
Matt Dilts, Héctor Muñoz-Avila
AAAI
2010
13 years 7 months ago
A Temporal Proof System for General Game Playing
A general game player is a system that understands the rules of unknown games and learns to play these games well without human intervention. A major challenge for research in Gen...
Michael Thielscher, Sebastian Voigt
ML
1998
ACM
136views Machine Learning» more  ML 1998»
13 years 5 months ago
Co-Evolution in the Successful Learning of Backgammon Strategy
Following Tesauro’s work on TD-Gammon, we used a 4000 parameter feed-forward neural network to develop a competitive backgammon evaluation function. Play proceeds by a roll of t...
Jordan B. Pollack, Alan D. Blair
ML
2002
ACM
154views Machine Learning» more  ML 2002»
13 years 5 months ago
Technical Update: Least-Squares Temporal Difference Learning
TD() is a popular family of algorithms for approximate policy evaluation in large MDPs. TD() works by incrementally updating the value function after each observed transition. It h...
Justin A. Boyan