Sciweavers

85 search results - page 3 / 17
» Solving Stochastic Planning Problems with Large State and Ac...
Sort
View
JAIR
2008
107views more  JAIR 2008»
15 years 1 months ago
Planning with Durative Actions in Stochastic Domains
Probabilistic planning problems are typically modeled as a Markov Decision Process (MDP). MDPs, while an otherwise expressive model, allow only for sequential, non-durative action...
Mausam, Daniel S. Weld
AIPS
1994
15 years 3 months ago
Solving Time-critical Decision-making Problems with Predictable Computational Demands
In this work we present an approach to solving time-critical decision-making problems by taking advantage of domain structure to expand the amountof time available for processing ...
Thomas Dean, Lloyd Greenwald
AAAI
2008
15 years 4 months ago
Limits and Possibilities of BDDs in State Space Search
This paper investigates the impact of symbolic search for solving domain-independent action planning problems with binary decision diagrams (BDDs). Polynomial upper and exponential...
Stefan Edelkamp, Peter Kissmann
ICTAI
2000
IEEE
15 years 5 months ago
Building efficient partial plans using Markov decision processes
Markov Decision Processes (MDP) have been widely used as a framework for planning under uncertainty. They allow to compute optimal sequences of actions in order to achieve a given...
Pierre Laroche
ICTAI
2005
IEEE
15 years 7 months ago
Planning with POMDPs Using a Compact, Logic-Based Representation
Partially Observable Markov Decision Processes (POMDPs) provide a general framework for AI planning, but they lack the structure for representing real world planning problems in a...
Chenggang Wang, James G. Schmolze