Sciweavers

83 search results - page 2 / 17
» Planning and Acting in Partially Observable Stochastic Domai...
Sort
View
AAAI
2010
13 years 6 months ago
Relational Partially Observable MDPs
Relational Markov Decision Processes (MDP) are a useraction for stochastic planning problems since one can develop abstract solutions for them that are independent of domain size ...
Chenggang Wang, Roni Khardon
JAIR
2000
152views more  JAIR 2000»
13 years 5 months ago
Value-Function Approximations for Partially Observable Markov Decision Processes
Partially observable Markov decision processes (POMDPs) provide an elegant mathematical framework for modeling complex decision and planning problems in stochastic domains in whic...
Milos Hauskrecht
TSMC
2010
13 years 7 hour ago
Active Learning of Plans for Safety and Reachability Goals With Partial Observability
Traditional planning assumes reachability goals and/or full observability. In this paper, we propose a novel solution for safety and reachability planning with partial observabilit...
Wonhong Nam, Rajeev Alur
AAAI
2010
13 years 6 months ago
PUMA: Planning Under Uncertainty with Macro-Actions
Planning in large, partially observable domains is challenging, especially when a long-horizon lookahead is necessary to obtain a good policy. Traditional POMDP planners that plan...
Ruijie He, Emma Brunskill, Nicholas Roy
ECML
2005
Springer
13 years 11 months ago
Using Rewards for Belief State Updates in Partially Observable Markov Decision Processes
Partially Observable Markov Decision Processes (POMDP) provide a standard framework for sequential decision making in stochastic environments. In this setting, an agent takes actio...
Masoumeh T. Izadi, Doina Precup