Sciweavers

334 search results - page 54 / 67
» How to Dynamically Merge Markov Decision Processes
Sort
View
JMLR
2006
190views more  JMLR 2006»
14 years 9 months ago
Causal Graph Based Decomposition of Factored MDPs
We present Variable Influence Structure Analysis, or VISA, an algorithm that performs hierarchical decomposition of factored Markov decision processes. VISA uses a dynamic Bayesia...
Anders Jonsson, Andrew G. Barto
PKDD
2010
Springer
129views Data Mining» more  PKDD 2010»
14 years 8 months ago
Smarter Sampling in Model-Based Bayesian Reinforcement Learning
Abstract. Bayesian reinforcement learning (RL) is aimed at making more efficient use of data samples, but typically uses significantly more computation. For discrete Markov Decis...
Pablo Samuel Castro, Doina Precup
GLOBECOM
2010
IEEE
14 years 7 months ago
Need-Based Communication for Smart Grid: When to Inquire Power Price?
In smart grid, a home appliance can adjust its power consumption level according to the realtime power price obtained from communication channels. Most studies on smart grid do not...
Husheng Li, Robert C. Qiu
GECCO
2009
Springer
162views Optimization» more  GECCO 2009»
14 years 7 months ago
Uncertainty handling CMA-ES for reinforcement learning
The covariance matrix adaptation evolution strategy (CMAES) has proven to be a powerful method for reinforcement learning (RL). Recently, the CMA-ES has been augmented with an ada...
Verena Heidrich-Meisner, Christian Igel
73
Voted
ICRA
2008
IEEE
128views Robotics» more  ICRA 2008»
15 years 4 months ago
A point-based POMDP planner for target tracking
— Target tracking has two variants that are often studied independently with different approaches: target searching requires a robot to find a target initially not visible, and ...
David Hsu, Wee Sun Lee, Nan Rong