Sciweavers

771 search results - page 7 / 155
» Markov Decision Processes with Arbitrary Reward Processes
Sort
View
ICML
2004
IEEE
16 years 15 days ago
Utile distinction hidden Markov models
This paper addresses the problem of constructing good action selection policies for agents acting in partially observable environments, a class of problems generally known as Part...
Daan Wierstra, Marco Wiering
ICML
2006
IEEE
16 years 15 days ago
An intrinsic reward mechanism for efficient exploration
How should a reinforcement learning agent act if its sole purpose is to efficiently learn an optimal policy for later use? In other words, how should it explore, to be able to exp...
Özgür Simsek, Andrew G. Barto
IROS
2009
IEEE
206views Robotics» more  IROS 2009»
15 years 6 months ago
Bayesian reinforcement learning in continuous POMDPs with gaussian processes
— Partially Observable Markov Decision Processes (POMDPs) provide a rich mathematical model to handle realworld sequential decision processes but require a known model to be solv...
Patrick Dallaire, Camille Besse, Stéphane R...
ML
2002
ACM
143views Machine Learning» more  ML 2002»
14 years 11 months ago
A Sparse Sampling Algorithm for Near-Optimal Planning in Large Markov Decision Processes
An issue that is critical for the application of Markov decision processes MDPs to realistic problems is how the complexity of planning scales with the size of the MDP. In stochas...
Michael J. Kearns, Yishay Mansour, Andrew Y. Ng
AAAI
1997
15 years 1 months ago
Structured Solution Methods for Non-Markovian Decision Processes
Markov Decision Processes (MDPs), currently a popular method for modeling and solving decision theoretic planning problems, are limited by the Markovian assumption: rewards and dy...
Fahiem Bacchus, Craig Boutilier, Adam J. Grove