Sciweavers

Share
ECML
2007
Springer

Efficient Continuous-Time Reinforcement Learning with Adaptive State Graphs

11 years 9 months ago
Efficient Continuous-Time Reinforcement Learning with Adaptive State Graphs
Abstract. We present a new reinforcement learning approach for deterministic continuous control problems in environments with unknown, arbitrary reward functions. The difficulty of finding solution trajectories for such problems can be reduced by incorporating limited prior knowledge of the approximative local system dynamics. The presented algorithm builds an adaptive state graph of sample points within the continuous state space. The nodes of the graph are generated by an efficient principled exploration scheme that directs the agent towards promising regions, while maintaining good online performance. Global solution trajectories are formed as combinations of local controllers that connect nodes of the graph, thereby naturally allowing continuous actions and continuous time steps. We demonstrate our approach on various movement planning tasks in continuous domains.
Gerhard Neumann, Michael Pfeiffer, Wolfgang Maass
Added 14 Aug 2010
Updated 14 Aug 2010
Type Conference
Year 2007
Where ECML
Authors Gerhard Neumann, Michael Pfeiffer, Wolfgang Maass
Comments (0)
books