Sciweavers

238 search results - page 3 / 48
» Value-Function Approximations for Partially Observable Marko...
Sort
View
CDC
2009
IEEE
134views Control Systems» more  CDC 2009»
13 years 9 months ago
Event-based control using quadratic approximate value functions
Abstract— In this paper we consider several problems involving control with limited actuation and sampling rates. Event-based control has emerged as an attractive approach for ad...
Randy Cogill
IJCAI
2001
13 years 6 months ago
An Improved Grid-Based Approximation Algorithm for POMDPs
Although a partially observable Markov decision process (POMDP) provides an appealing model for problems of planning under uncertainty, exact algorithms for POMDPs are intractable...
Rong Zhou, Eric A. Hansen
ICML
2007
IEEE
14 years 5 months ago
Constructing basis functions from directed graphs for value function approximation
Basis functions derived from an undirected graph connecting nearby samples from a Markov decision process (MDP) have proven useful for approximating value functions. The success o...
Jeffrey Johns, Sridhar Mahadevan
ICRA
2007
IEEE
155views Robotics» more  ICRA 2007»
13 years 11 months ago
Value Function Approximation on Non-Linear Manifolds for Robot Motor Control
— The least squares approach works efficiently in value function approximation, given appropriate basis functions. Because of its smoothness, the Gaussian kernel is a popular an...
Masashi Sugiyama, Hirotaka Hachiya, Christopher To...
NN
2010
Springer
187views Neural Networks» more  NN 2010»
12 years 11 months ago
Efficient exploration through active learning for value function approximation in reinforcement learning
Appropriately designing sampling policies is highly important for obtaining better control policies in reinforcement learning. In this paper, we first show that the least-squares ...
Takayuki Akiyama, Hirotaka Hachiya, Masashi Sugiya...