Sciweavers

238 search results - page 17 / 48
» Value-Function Approximations for Partially Observable Marko...
Sort
View
IROS
2009
IEEE
206views Robotics» more  IROS 2009»
15 years 6 months ago
Bayesian reinforcement learning in continuous POMDPs with gaussian processes
— Partially Observable Markov Decision Processes (POMDPs) provide a rich mathematical model to handle realworld sequential decision processes but require a known model to be solv...
Patrick Dallaire, Camille Besse, Stéphane R...
PKDD
2010
Springer
164views Data Mining» more  PKDD 2010»
14 years 9 months ago
Efficient Planning in Large POMDPs through Policy Graph Based Factorized Approximations
Partially observable Markov decision processes (POMDPs) are widely used for planning under uncertainty. In many applications, the huge size of the POMDP state space makes straightf...
Joni Pajarinen, Jaakko Peltonen, Ari Hottinen, Mik...
ATAL
2003
Springer
15 years 4 months ago
Transition-independent decentralized markov decision processes
There has been substantial progress with formal models for sequential decision making by individual agents using the Markov decision process (MDP). However, similar treatment of m...
Raphen Becker, Shlomo Zilberstein, Victor R. Lesse...
ICC
2007
IEEE
137views Communications» more  ICC 2007»
15 years 6 months ago
Optimality and Complexity of Opportunistic Spectrum Access: A Truncated Markov Decision Process Formulation
— We consider opportunistic spectrum access (OSA) which allows secondary users to identify and exploit instantaneous spectrum opportunities resulting from the bursty traffic of ...
Dejan V. Djonin, Qing Zhao, Vikram Krishnamurthy
ICML
2008
IEEE
16 years 14 days ago
Reinforcement learning with limited reinforcement: using Bayes risk for active learning in POMDPs
Partially Observable Markov Decision Processes (POMDPs) have succeeded in planning domains that require balancing actions that increase an agent's knowledge and actions that ...
Finale Doshi, Joelle Pineau, Nicholas Roy