Sciweavers

28 search results - page 1 / 6
» Policy Improvement for POMDPs Using Normalized Importance Sa...
Sort
View
UAI
2001
13 years 6 months ago
Policy Improvement for POMDPs Using Normalized Importance Sampling
We present a new method for estimating the expected return of a POMDP from experience. The estimator does not assume any knowledge of the POMDP, can estimate the returns for finit...
Christian R. Shelton
AAAI
2006
13 years 6 months ago
Incremental Least Squares Policy Iteration for POMDPs
We present a new algorithm, called incremental least squares policy iteration (ILSPI), for finding the infinite-horizon stationary policy for partially observable Markov decision ...
Hui Li, Xuejun Liao, Lawrence Carin
ATAL
2006
Springer
13 years 8 months ago
Solving POMDPs using quadratically constrained linear programs
Developing scalable algorithms for solving partially observable Markov decision processes (POMDPs) is an important challenge. One promising approach is based on representing POMDP...
Christopher Amato, Daniel S. Bernstein, Shlomo Zil...
AAAI
2008
13 years 7 months ago
A Variance Analysis for POMDP Policy Evaluation
Partially Observable Markov Decision Processes have been studied widely as a model for decision making under uncertainty, and a number of methods have been developed to find the s...
Mahdi Milani Fard, Joelle Pineau, Peng Sun
ATAL
2008
Springer
13 years 7 months ago
Value-based observation compression for DEC-POMDPs
Representing agent policies compactly is essential for improving the scalability of multi-agent planning algorithms. In this paper, we focus on developing a pruning technique that...
Alan Carlin, Shlomo Zilberstein