Parallel Reinforcement Learning with Linear Function Approximation

9 years 9 months ago
Parallel Reinforcement Learning with Linear Function Approximation
In this paper, we investigate the use of parallelization in reinforcement learning (RL), with the goal of learning optimal policies for single-agent RL problems more quickly by using parallel hardware. Our approach is based on agents using the SARSA(λ) algorithm, with value functions represented using linear function approximators. In our proposed method, each agent learns independently in a separate simulation of the single-agent problem. The agents periodically exchange information extracted from the weights of their approximators, accelerating convergence towards the optimal policy. We present empirical results for an implementation on a Beowulf cluster. Categories and Subject Descriptors I.2.6 [Artificial Intelligence]: Learning General Terms Algorithms, Performance, Experimentation Keywords Reinforcement learning, value function approximation, parallel algorithms
Matthew Grounds, Daniel Kudenko
Added 08 Dec 2010
Updated 08 Dec 2010
Type Journal
Year 2007
Authors Matthew Grounds, Daniel Kudenko
Comments (0)