Parallel reinforcement learning with linear function approximation

Matthew Grounds, Daniel Kudenko

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

In this paper, we investigate the use of parallelization in reinforcement learning (RL), with the goal of learning optimal policies for single-agent RL problems more quickly by using parallel hardware. Our approach is based on agents using the SARSA(lambda) algorithm, with value functions represented using, linear function approximators. In our proposed method, each agent learns independently in a separate simulation of the single-agent problem. The agents periodically exchange information extracted from the weights of their approximators, accelerating convergence towards the optimal policy. We develop three increasingly efficient versions of this approach to parallel RL, and present empirical results for an implementation of the methods on a Beowulf cluster.

Original languageEnglish
Title of host publicationADAPTIVE AGENTS AND MULTI-AGENT SYSTEMS
EditorsK Tuyls, A Nowe, Z Guessoum, D Kudenko
Place of PublicationBERLIN
PublisherSpringer
Pages60-74
Number of pages15
Volume4865 LNAI
ISBN (Print)978-3-540-77947-6
Publication statusPublished - 2008

Cite this