Sciweavers

36 search results - page 3 / 8
» Using Linear Programming for Bayesian Exploration in Markov ...
Sort
View
ICML
2006
IEEE
14 years 11 days ago
Automatic basis function construction for approximate dynamic programming and reinforcement learning
We address the problem of automatically constructing basis functions for linear approximation of the value function of a Markov Decision Process (MDP). Our work builds on results ...
Philipp W. Keller, Shie Mannor, Doina Precup
ATAL
2006
Springer
13 years 10 months ago
Solving POMDPs using quadratically constrained linear programs
Developing scalable algorithms for solving partially observable Markov decision processes (POMDPs) is an important challenge. One promising approach is based on representing POMDP...
Christopher Amato, Daniel S. Bernstein, Shlomo Zil...
UAI
2004
13 years 7 months ago
Solving Factored MDPs with Continuous and Discrete Variables
Although many real-world stochastic planning problems are more naturally formulated by hybrid models with both discrete and continuous variables, current state-of-the-art methods ...
Carlos Guestrin, Milos Hauskrecht, Branislav Kveto...
AAAI
2011
12 years 6 months ago
Linear Dynamic Programs for Resource Management
Sustainable resource management in many domains presents large continuous stochastic optimization problems, which can often be modeled as Markov decision processes (MDPs). To solv...
Marek Petrik, Shlomo Zilberstein
DATE
2008
IEEE
136views Hardware» more  DATE 2008»
14 years 26 days ago
A Framework of Stochastic Power Management Using Hidden Markov Model
- The effectiveness of stochastic power management relies on the accurate system and workload model and effective policy optimization. Workload modeling is a machine learning proce...
Ying Tan, Qinru Qiu