Sciweavers

13 search results - page 2 / 3
» Minimal Sufficient Explanations for Factored Markov Decision...
Sort
View
PKDD
2010
Springer
164views Data Mining» more  PKDD 2010»
13 years 3 months ago
Efficient Planning in Large POMDPs through Policy Graph Based Factorized Approximations
Partially observable Markov decision processes (POMDPs) are widely used for planning under uncertainty. In many applications, the huge size of the POMDP state space makes straightf...
Joni Pajarinen, Jaakko Peltonen, Ari Hottinen, Mik...
ICTAI
2010
IEEE
13 years 3 months ago
A Closer Look at MOMDPs
Abstract--The difficulties encountered in sequential decisionmaking problems under uncertainty are often linked to the large size of the state space. Exploiting the structure of th...
Mauricio Araya-López, Vincent Thomas, Olivi...
IUI
2010
ACM
14 years 2 months ago
A POMDP approach to P300-based brain-computer interfaces
Most of the previous work on non-invasive brain-computer interfaces (BCIs) has been focused on feature extraction and classification algorithms to achieve high performance for the...
Jaeyoung Park, Kee-Eung Kim, Sungho Jo
NIPS
2000
13 years 6 months ago
Using Free Energies to Represent Q-values in a Multiagent Reinforcement Learning Task
The problem of reinforcement learning in large factored Markov decision processes is explored. The Q-value of a state-action pair is approximated by the free energy of a product o...
Brian Sallans, Geoffrey E. Hinton
CORR
2006
Springer
113views Education» more  CORR 2006»
13 years 5 months ago
A Unified View of TD Algorithms; Introducing Full-Gradient TD and Equi-Gradient Descent TD
This paper addresses the issue of policy evaluation in Markov Decision Processes, using linear function approximation. It provides a unified view of algorithms such as TD(), LSTD()...
Manuel Loth, Philippe Preux