Sciweavers

280 search results - page 22 / 56
» Planning for Markov Decision Processes with Sparse Stochasti...
Sort
View
AAAI
2000
14 years 11 months ago
Decision Making under Uncertainty: Operations Research Meets AI (Again)
Models for sequential decision making under uncertainty (e.g., Markov decision processes,or MDPs) have beenstudied in operations research for decades. The recent incorporation of ...
Craig Boutilier
IJCAI
2003
14 years 11 months ago
Automated Generation of Understandable Contingency Plans
Markov decision processes (MDPs) and contingency planning (CP) are two widely used approaches to planning under uncertainty. MDPs are attractive because the model is extremely gen...
Max Horstmann, Shlomo Zilberstein
FLAIRS
2001
14 years 11 months ago
Probabilistic Planning for Behavior-Based Robots
Partially Observable Markov Decision Process models (POMDPs) have been applied to low-level robot control. We show how to use POMDPs differently, namely for sensorplanning in the ...
Amin Atrash, Sven Koenig
ACL
2000
14 years 11 months ago
Spoken Dialogue Management Using Probabilistic Reasoning
Spoken dialogue managers have benefited from using stochastic planners such as Markov Decision Processes (MDPs). However, so far, MDPs do not handle well noisy and ambiguous speec...
Nicholas Roy, Joelle Pineau, Sebastian Thrun
AIED
2011
Springer
14 years 1 months ago
Faster Teaching by POMDP Planning
Both human and automated tutors must infer what a student knows and plan future actions to maximize learning. Though substantial research has been done on tracking and modeling stu...
Anna N. Rafferty, Emma Brunskill, Thomas L. Griffi...