Sciweavers

6 search results - page 1 / 2
» Optimally Solving Dec-POMDPs as Continuous-State MDPs
Sort
View
JAIR
2008
126views more  JAIR 2008»
13 years 5 months ago
Optimal and Approximate Q-value Functions for Decentralized POMDPs
Decision-theoretic planning is a popular approach to sequential decision making problems, because it treats uncertainty in sensing and acting in a principled way. In single-agent ...
Frans A. Oliehoek, Matthijs T. J. Spaan, Nikos A. ...
AIPS
2006
13 years 6 months ago
Solving Factored MDPs with Exponential-Family Transition Models
Markov decision processes (MDPs) with discrete and continuous state and action components can be solved efficiently by hybrid approximate linear programming (HALP). The main idea ...
Branislav Kveton, Milos Hauskrecht
ICML
2010
IEEE
13 years 6 months ago
Inverse Optimal Control with Linearly-Solvable MDPs
We present new algorithms for inverse optimal control (or inverse reinforcement learning, IRL) within the framework of linearlysolvable MDPs (LMDPs). Unlike most prior IRL algorit...
Dvijotham Krishnamurthy, Emanuel Todorov
AAAI
2006
13 years 6 months ago
Learning Basis Functions in Hybrid Domains
Markov decision processes (MDPs) with discrete and continuous state and action components can be solved efficiently by hybrid approximate linear programming (HALP). The main idea ...
Branislav Kveton, Milos Hauskrecht
AAAI
2012
11 years 7 months ago
Kernel-Based Reinforcement Learning on Representative States
Markov decision processes (MDPs) are an established framework for solving sequential decision-making problems under uncertainty. In this work, we propose a new method for batchmod...
Branislav Kveton, Georgios Theocharous