Sciweavers

201 search results - page 8 / 41
» Solving Concurrent Markov Decision Processes
Sort
View
ICRA
2007
IEEE
154views Robotics» more  ICRA 2007»
15 years 6 months ago
Oracular Partially Observable Markov Decision Processes: A Very Special Case
— We introduce the Oracular Partially Observable Markov Decision Process (OPOMDP), a type of POMDP in which the world produces no observations; instead there is an “oracle,” ...
Nicholas Armstrong-Crews, Manuela M. Veloso
IJCAI
2007
15 years 1 months ago
Using Linear Programming for Bayesian Exploration in Markov Decision Processes
A key problem in reinforcement learning is finding a good balance between the need to explore the environment and the need to gain rewards by exploiting existing knowledge. Much ...
Pablo Samuel Castro, Doina Precup
UAI
1998
15 years 1 months ago
Structured Reachability Analysis for Markov Decision Processes
Recent research in decision theoretic planning has focussedon making the solution of Markov decision processes (MDPs) more feasible. We develop a family of algorithms for structur...
Craig Boutilier, Ronen I. Brafman, Christopher W. ...
ICTAI
2000
IEEE
15 years 3 months ago
Building efficient partial plans using Markov decision processes
Markov Decision Processes (MDP) have been widely used as a framework for planning under uncertainty. They allow to compute optimal sequences of actions in order to achieve a given...
Pierre Laroche
CDC
2009
IEEE
169views Control Systems» more  CDC 2009»
15 years 4 months ago
Parametric regret in uncertain Markov decision processes
— We consider decision making in a Markovian setup where the reward parameters are not known in advance. Our performance criterion is the gap between the performance of the best ...
Huan Xu, Shie Mannor