Title

Scalable Multiagent Learning Through Indirect Encoding Of Policy Geometry

Keywords

HyperNEAT; Indirect encoding; Multiagent learning; Neural networks

Abstract

Multiagent systems present many challenging, real-world problems to artificial intelligence. Because it is difficult to engineer the behaviors of multiple cooperating agents by hand, multiagent learning has become a popular approach to their design. While there are a variety of traditional approaches to multiagent learning, many suffer from increased computational costs for large teams and the problem of reinvention (that is, the inability to recognize that certain skills are shared by some or all team member). This paper presents an alternative approach to multiagent learning called multiagent HyperNEAT that represents the team as a pattern of policies rather than as a set of individual agents. The main idea is that an agent's location within a canonical team layout (which can be physical, such as positions on a sports team, or conceptual, such as an agent's relative speed) tends to dictate its role within that team. This paper introduces the term policy geometry to describe this relationship between role and position on the team. Interestingly, such patterns effectively represent up to an infinite number of multiagent policies that can be sampled from the policy geometry as needed to allow training very large teams or, in some cases, scaling up the size of a team without additional learning. In this paper, multiagent HyperNEAT is compared to a traditional learning method, multiagent Sarsa(λ), in a predator-prey domain, where it demonstrates its ability to train large teams. © 2013 Springer-Verlag Berlin Heidelberg.

Publication Date

6-1-2013

Publication Title

Evolutionary Intelligence

Volume

6

Issue

1

Number of Pages

1-26

Document Type

Article

Personal Identifier

scopus

DOI Link

https://doi.org/10.1007/s12065-012-0086-3

Socpus ID

84872202698 (Scopus)

Source API URL

https://api.elsevier.com/content/abstract/scopus_id/84872202698

This document is currently not available here.

Share

COinS