Sciweavers

36 search results - page 1 / 8
» Using Linear Programming for Bayesian Exploration in Markov ...
Sort
View
IJCAI
2007
13 years 6 months ago
Using Linear Programming for Bayesian Exploration in Markov Decision Processes
A key problem in reinforcement learning is finding a good balance between the need to explore the environment and the need to gain rewards by exploiting existing knowledge. Much ...
Pablo Samuel Castro, Doina Precup
SARA
2007
Springer
13 years 11 months ago
Active Learning of Dynamic Bayesian Networks in Markov Decision Processes
Several recent techniques for solving Markov decision processes use dynamic Bayesian networks to compactly represent tasks. The dynamic Bayesian network representation may not be g...
Anders Jonsson, Andrew G. Barto
ICML
2010
IEEE
13 years 6 months ago
Feature Selection Using Regularization in Approximate Linear Programs for Markov Decision Processes
Approximate dynamic programming has been used successfully in a large variety of domains, but it relies on a small set of provided approximation features to calculate solutions re...
Marek Petrik, Gavin Taylor, Ronald Parr, Shlomo Zi...
UAI
2004
13 years 6 months ago
Dynamic Programming for Structured Continuous Markov Decision Problems
We describe an approach for exploiting structure in Markov Decision Processes with continuous state variables. At each step of the dynamic programming, the state space is dynamica...
Zhengzhu Feng, Richard Dearden, Nicolas Meuleau, R...
ECAI
2010
Springer
13 years 6 months ago
On Finding Compromise Solutions in Multiobjective Markov Decision Processes
A Markov Decision Process (MDP) is a general model for solving planning problems under uncertainty. It has been extended to multiobjective MDP to address multicriteria or multiagen...
Patrice Perny, Paul Weng