In this work we extend the work of Dean, Kaelbling, Kirman and Nicholson on planning under time constraints in stochastic domains to handle more complicated scheduling problems. I...
The ways in which an agent’s actions affect the world can often be modeled compactly using a set of relational probabilistic planning rules. This paper addresses the problem of ...
Ashwin Deshpande, Brian Milch, Luke S. Zettlemoyer...
ys when planning meant searching for a sequence of abstract actions that satisfied some symbolic predicate. Robots can now learn their own representations through statistical infe...
Partially Observable Markov Decision Processes (POMDP) provide a standard framework for sequential decision making in stochastic environments. In this setting, an agent takes actio...
Relational Markov Decision Processes (MDP) are a useraction for stochastic planning problems since one can develop abstract solutions for them that are independent of domain size ...