Sciweavers

201 search results - page 23 / 41
» Solving Concurrent Markov Decision Processes
Sort
View
AAAI
2010
14 years 11 months ago
Relational Partially Observable MDPs
Relational Markov Decision Processes (MDP) are a useraction for stochastic planning problems since one can develop abstract solutions for them that are independent of domain size ...
Chenggang Wang, Roni Khardon
JAIR
2008
130views more  JAIR 2008»
14 years 9 months ago
Online Planning Algorithms for POMDPs
Partially Observable Markov Decision Processes (POMDPs) provide a rich framework for sequential decision-making under uncertainty in stochastic domains. However, solving a POMDP i...
Stéphane Ross, Joelle Pineau, Sébast...
CDC
2008
IEEE
118views Control Systems» more  CDC 2008»
15 years 4 months ago
A density projection approach to dimension reduction for continuous-state POMDPs
Abstract— Research on numerical solution methods for partially observable Markov decision processes (POMDPs) has primarily focused on discrete-state models, and these algorithms ...
Enlu Zhou, Michael C. Fu, Steven I. Marcus
JMLR
2010
125views more  JMLR 2010»
14 years 4 months ago
Variational methods for Reinforcement Learning
We consider reinforcement learning as solving a Markov decision process with unknown transition distribution. Based on interaction with the environment, an estimate of the transit...
Thomas Furmston, David Barber
ATAL
2009
Springer
15 years 4 months ago
Planning with continuous resources for agent teams
Many problems of multiagent planning under uncertainty require distributed reasoning with continuous resources and resource limits. Decentralized Markov Decision Problems (Dec-MDP...
Janusz Marecki, Milind Tambe