Sciweavers

682 search results - page 120 / 137
» One-Counter Markov Decision Processes
Sort
View
107
Voted
AAAI
2007
15 years 3 months ago
Continuous State POMDPs for Object Manipulation Tasks
My research focus is on using continuous state partially observable Markov decision processes (POMDPs) to perform object manipulation tasks using a robotic arm. During object mani...
Emma Brunskill
124
Voted
AAAI
2007
15 years 3 months ago
Authorial Idioms for Target Distributions in TTD-MDPs
In designing Markov Decision Processes (MDP), one must define the world, its dynamics, a set of actions, and a reward function. MDPs are often applied in situations where there i...
David L. Roberts, Sooraj Bhat, Kenneth St. Clair, ...
124
Voted
AAAI
2008
15 years 3 months ago
Towards Faster Planning with Continuous Resources in Stochastic Domains
Agents often have to construct plans that obey resource limits for continuous resources whose consumption can only be characterized by probability distributions. While Markov Deci...
Janusz Marecki, Milind Tambe
89
Voted
AAAI
2007
15 years 3 months ago
Compact Spectral Bases for Value Function Approximation Using Kronecker Factorization
A new spectral approach to value function approximation has recently been proposed to automatically construct basis functions from samples. Global basis functions called proto-val...
Jeffrey Johns, Sridhar Mahadevan, Chang Wang
124
Voted
AAAI
2007
15 years 3 months ago
Thresholded Rewards: Acting Optimally in Timed, Zero-Sum Games
In timed, zero-sum games, the goal is to maximize the probability of winning, which is not necessarily the same as maximizing our expected reward. We consider cumulative intermedi...
Colin McMillen, Manuela M. Veloso