Multi-agent, reward shaping for RoboCup KeepAway

Sam Devlin, Marek Grześ, Daniel Kudenko

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

This paper investigates the impact of reward shaping in multi-agent reinforcement learning as a way to incorporate domain knowledge about good strategies. In theory [2], potential-based reward shaping does not alter the Nash Equilibria of a stochastic game, only the exploration of the shaped agent. We demonstrate empirically the performance of state-based and state-action-based reward shaping in RoboCup KeepAway. The results illustrate that reward shaping can alter both the learning time required to reach a stable joint policy and the final group performance for better or worse.

Original languageEnglish
Title of host publication10th International Conference on Autonomous Agents and Multiagent Systems 2011, AAMAS 2011
PublisherInternational Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS)
Pages1157-1158
Number of pages2
Volume2
Publication statusPublished - 2011
Event10th International Conference on Autonomous Agents and Multiagent Systems 2011, AAMAS 2011 - Taipei, Taiwan
Duration: 2 May 20116 May 2011

Conference

Conference10th International Conference on Autonomous Agents and Multiagent Systems 2011, AAMAS 2011
Country/TerritoryTaiwan
CityTaipei
Period2/05/116/05/11

Keywords

  • Multiagent Learning
  • Reinforcement Learning
  • Reward Shaping
  • Reward Structures for Learning

Cite this