Sciweavers

89 search results - page 13 / 18
» Sample-Based Planning for Continuous Action Markov Decision ...
Sort
View
JAIR
2008
130views more  JAIR 2008»
14 years 9 months ago
Online Planning Algorithms for POMDPs
Partially Observable Markov Decision Processes (POMDPs) provide a rich framework for sequential decision-making under uncertainty in stochastic domains. However, solving a POMDP i...
Stéphane Ross, Joelle Pineau, Sébast...
ICML
2006
IEEE
15 years 10 months ago
PAC model-free reinforcement learning
For a Markov Decision Process with finite state (size S) and action spaces (size A per state), we propose a new algorithm--Delayed Q-Learning. We prove it is PAC, achieving near o...
Alexander L. Strehl, Lihong Li, Eric Wiewiora, Joh...
AIPS
2003
14 years 11 months ago
A Framework for Planning in Continuous-time Stochastic Domains
We propose a framework for policy generation in continuoustime stochastic domains with concurrent actions and events of uncertain duration. We make no assumptions regarding the co...
Håkan L. S. Younes, David J. Musliner, Reid ...
NIPS
2008
14 years 11 months ago
Particle Filter-based Policy Gradient in POMDPs
Our setting is a Partially Observable Markov Decision Process with continuous state, observation and action spaces. Decisions are based on a Particle Filter for estimating the bel...
Pierre-Arnaud Coquelin, Romain Deguest, Rém...
HICSS
2003
IEEE
123views Biometrics» more  HICSS 2003»
15 years 2 months ago
Issues in Rational Planning in Multi-Agent Settings
We adopt the decision-theoretic principle of expected utility maximization as a paradigm for designing autonomous rational agents operating in multi-agent environments. We use the...
Piotr J. Gmytrasiewicz