Title
Task Switching In Multirobot Learning Through Indirect Encoding
Abstract
Multirobot domains are a challenge for learning algorithms because they require robots to learn to cooperate to achieve a common goal. The challenge only becomes greater when robots must perform heterogeneous tasks to reach that goal. Multiagent HyperNEAT is a neuroevolutionary method (i.e. a method that evolves neural networks) that has proven successful in several cooperative multiagent domains by exploiting the concept of policy geometry, which means the policies of team members are learned as a function of how they relate to each other based on canonical starting positions. This paper extends the multiagent HyperNEAT algorithm by introducing situational policy geometry, which allows each agent to encode multiple policies that can be switched depending on the agent's state. This concept is demonstrated both in simulation and in real Khepera III robots in a patrol and return task, where robots must cooperate to cover an area and return home when called. Robot teams that are trained with situational policy geometry are compared to teams that are not and shown to find solutions more consistently that are also able to transfer to the real world. © 2011 IEEE.
Publication Date
12-29-2011
Publication Title
IEEE International Conference on Intelligent Robots and Systems
Number of Pages
2802-2809
Document Type
Article; Proceedings Paper
Personal Identifier
scopus
DOI Link
https://doi.org/10.1109/IROS.2011.6048150
Copyright Status
Unknown
Socpus ID
84455207473 (Scopus)
Source API URL
https://api.elsevier.com/content/abstract/scopus_id/84455207473
STARS Citation
D'Ambrosio, David B.; Lehman, Joel; Risi, Sebastian; and Stanley, Kenneth O., "Task Switching In Multirobot Learning Through Indirect Encoding" (2011). Scopus Export 2010-2014. 2399.
https://stars.library.ucf.edu/scopus2010/2399