Title

Task Switching In Multirobot Learning Through Indirect Encoding

Abstract

Multirobot domains are a challenge for learning algorithms because they require robots to learn to cooperate to achieve a common goal. The challenge only becomes greater when robots must perform heterogeneous tasks to reach that goal. Multiagent HyperNEAT is a neuroevolutionary method (i.e. a method that evolves neural networks) that has proven successful in several cooperative multiagent domains by exploiting the concept of policy geometry, which means the policies of team members are learned as a function of how they relate to each other based on canonical starting positions. This paper extends the multiagent HyperNEAT algorithm by introducing situational policy geometry, which allows each agent to encode multiple policies that can be switched depending on the agent's state. This concept is demonstrated both in simulation and in real Khepera III robots in a patrol and return task, where robots must cooperate to cover an area and return home when called. Robot teams that are trained with situational policy geometry are compared to teams that are not and shown to find solutions more consistently that are also able to transfer to the real world. © 2011 IEEE.

Publication Date

12-29-2011

Publication Title

IEEE International Conference on Intelligent Robots and Systems

Number of Pages

2802-2809

Document Type

Article; Proceedings Paper

Personal Identifier

scopus

DOI Link

https://doi.org/10.1109/IROS.2011.6048150

Socpus ID

84455207473 (Scopus)

Source API URL

https://api.elsevier.com/content/abstract/scopus_id/84455207473

This document is currently not available here.

Share

COinS