Sciweavers

48 search results - page 3 / 10
» Approximate Planning in POMDPs with Macro-Actions
Sort
View
ICML
2008
IEEE
14 years 6 months ago
Reinforcement learning with limited reinforcement: using Bayes risk for active learning in POMDPs
Partially Observable Markov Decision Processes (POMDPs) have succeeded in planning domains that require balancing actions that increase an agent's knowledge and actions that ...
Finale Doshi, Joelle Pineau, Nicholas Roy
ATAL
2008
Springer
13 years 8 months ago
The permutable POMDP: fast solutions to POMDPs for preference elicitation
The ability for an agent to reason under uncertainty is crucial for many planning applications, since an agent rarely has access to complete, error-free information about its envi...
Finale Doshi, Nicholas Roy
ATAL
2010
Springer
13 years 7 months ago
Risk-sensitive planning in partially observable environments
Partially Observable Markov Decision Process (POMDP) is a popular framework for planning under uncertainty in partially observable domains. Yet, the POMDP model is riskneutral in ...
Janusz Marecki, Pradeep Varakantham
AAAI
2006
13 years 7 months ago
Compact, Convex Upper Bound Iteration for Approximate POMDP Planning
Partially observable Markov decision processes (POMDPs) are an intuitive and general way to model sequential decision making problems under uncertainty. Unfortunately, even approx...
Tao Wang, Pascal Poupart, Michael H. Bowling, Dale...
JAIR
2000
152views more  JAIR 2000»
13 years 5 months ago
Value-Function Approximations for Partially Observable Markov Decision Processes
Partially observable Markov decision processes (POMDPs) provide an elegant mathematical framework for modeling complex decision and planning problems in stochastic domains in whic...
Milos Hauskrecht