Sciweavers

48 search results - page 6 / 10
» Approximate Planning in POMDPs with Macro-Actions
Sort
View
94
Voted
ATAL
2010
Springer
14 years 10 months ago
Point-based backup for decentralized POMDPs: complexity and new algorithms
Decentralized POMDPs provide an expressive framework for sequential multi-agent decision making. Despite their high complexity, there has been significant progress in scaling up e...
Akshat Kumar, Shlomo Zilberstein
ATAL
2009
Springer
15 years 4 months ago
Lossless clustering of histories in decentralized POMDPs
Decentralized partially observable Markov decision processes (Dec-POMDPs) constitute a generic and expressive framework for multiagent planning under uncertainty. However, plannin...
Frans A. Oliehoek, Shimon Whiteson, Matthijs T. J....
AAAI
1996
14 years 10 months ago
Computing Optimal Policies for Partially Observable Decision Processes Using Compact Representations
: Partially-observable Markov decision processes provide a very general model for decision-theoretic planning problems, allowing the trade-offs between various courses of actions t...
Craig Boutilier, David Poole
ICASSP
2011
IEEE
14 years 1 months ago
Bayesian reinforcement learning for POMDP-based dialogue systems
Spoken dialogue systems are gaining popularity with improvements in speech recognition technologies. Dialogue systems can be modeled effectively using POMDPs, achieving improvemen...
ShaoWei Png, Joelle Pineau
ATAL
2007
Springer
15 years 3 months ago
Q-value functions for decentralized POMDPs
Planning in single-agent models like MDPs and POMDPs can be carried out by resorting to Q-value functions: a (near-) optimal Q-value function is computed in a recursive manner by ...
Frans A. Oliehoek, Nikos A. Vlassis