Sciweavers

12 search results - page 1 / 3
» Monte Carlo Value Iteration for Continuous-State POMDPs
Sort
View

Publication
273views
12 years 11 months ago
Monte Carlo Value Iteration for Continuous-State POMDPs
Partially observable Markov decision processes (POMDPs) have been successfully applied to various robot motion planning tasks under uncertainty. However, most existing POMDP algo...
Haoyu Bai, David Hsu, Wee Sun Lee, and Vien A. Ngo
JMLR
2006
116views more  JMLR 2006»
13 years 4 months ago
Point-Based Value Iteration for Continuous POMDPs
We propose a novel approach to optimize Partially Observable Markov Decisions Processes (POMDPs) defined on continuous spaces. To date, most algorithms for model-based POMDPs are ...
Josep M. Porta, Nikos A. Vlassis, Matthijs T. J. S...
NIPS
2007
13 years 6 months ago
Reinforcement Learning in Continuous Action Spaces through Sequential Monte Carlo Methods
Learning in real-world domains often requires to deal with continuous state and action spaces. Although many solutions have been proposed to apply Reinforcement Learning algorithm...
Alessandro Lazaric, Marcello Restelli, Andrea Bona...
ICML
1999
IEEE
14 years 5 months ago
Monte Carlo Hidden Markov Models: Learning Non-Parametric Models of Partially Observable Stochastic Processes
We present a learning algorithm for non-parametric hidden Markov models with continuous state and observation spaces. All necessary probability densities are approximated using sa...
Sebastian Thrun, John Langford, Dieter Fox
ICML
2005
IEEE
14 years 5 months ago
Finite time bounds for sampling based fitted value iteration
In this paper we consider sampling based fitted value iteration for discounted, large (possibly infinite) state space, finite action Markovian Decision Problems where only a gener...
Csaba Szepesvári, Rémi Munos