Sciweavers

Share
AIPS
2009

Incremental Policy Generation for Finite-Horizon DEC-POMDPs

10 years 19 days ago
Incremental Policy Generation for Finite-Horizon DEC-POMDPs
Solving multiagent planning problems modeled as DECPOMDPs is an important challenge. These models are often solved by using dynamic programming, but the high resource usage of current approaches results in limited scalability. To improve the efficiency of dynamic programming algorithms, we propose a new backup algorithm that is based on a reachability analysis of the state space. This method, which we call incremental policy generation, can be used to produce an optimal solution for any possible initial state or further scalability can be achieved by making use of a known start state. When incorporated into the optimal dynamic programming algorithm, our experiments show that planning horizon can be increased due to a marked reduction in resource consumption. This approach also fits nicely with approximate dynamic programming algorithms. To demonstrate this, we incorporate it into the state-of-the-art PBIP algorithm and show significant performance gains. The results suggest that the p...
Christopher Amato, Jilles Steeve Dibangoye, Shlomo
Added 08 Nov 2010
Updated 08 Nov 2010
Type Conference
Year 2009
Where AIPS
Authors Christopher Amato, Jilles Steeve Dibangoye, Shlomo Zilberstein
Comments (0)
books