IPCAT 2015: Information Processing in Cells and Tissues pp 75-89 | Cite as
Team Search Tactics Through Multi-Agent HyperNEAT
Conference paper
First Online:
Abstract
User defined tactics for teams of unmanned systems can be brittle and difficult to define. The state and action space grows with each new system added to the team which increases the difficultly in designing robust behaviors. In this paper we present a method for using Multi-agent HyperNEAT to develop tactics for a team of simulated unmanned systems that is robust to novel situations, and scales with the number of team members. We focus on the tactics of a search area coverage task, where the need for team work, and robust asset management are critical to success.
Keywords
Search Task Search Pattern Relative Sensor Radar Sensor Radar Segment
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
- 1.Aaltonen, T., Adelman, J., Akimoto, T., Albrow, M.G., González, B.Á., Amerio, S., Amidei, D., Anastassov, A.: Measurement of the top-quark mass with dilepton events selected using neuroevolution at CDF. Phys. Rev. Lett. 102, 152001 (2009)CrossRefGoogle Scholar
- 2.Cardamone, L., Loiacono, D., Lanzi, P.L.: Learning to drive in the open racing car simulator using online neuroevolution. IEEE Trans. Comput. Intell. AI Games 2, 176–190 (2010)CrossRefGoogle Scholar
- 3.D’Ambrosio, D.B., Lehman, J., Risi, S.: MA-HyperNEAT AgentSimulator. http://eplex.cs.ucf.edu/software/AgentSimulator_v1_0.zip
- 4.D’Ambrosio, D.B., Lehman, J., Risi, S., Stanley, K.O.: Evolving policy geometry for scalable multiagent learning. In: Proceedings of the Ninth International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2010), pp. 731–738 (2010)Google Scholar
- 5.D’Ambrosio, D.B., Stanley, K.O.: Generative encoding for multiagent learning. In: Proceedings of the 10th Annual Conference on Genetic and Evolutionary Computation GECCO 2008, vol. 2008, p. 819 (2008)Google Scholar
- 6.Dragoni, A.F., Giorgini, P.: Belief revision through the belief-function formalism in a multi-agent environment. In: Müller, J.P., Wooldridge, M.J., Jennings, N.R. (eds.) Intelligent Agents III Agent Theories, Architectures, and Languages. LNCS, vol. 1193, pp. 103–115. Springer, Heidelberg (1997)CrossRefGoogle Scholar
- 7.D’Ambrosio, D.B., Stanley, K.O.: Scalable multiagent learning through indirect encoding of policy geometry. Evol. Intel. 6(1), 1–26 (2013)CrossRefGoogle Scholar
- 8.Friedrich, H., Rogalla, O., Dillmann, R.: Communication and propagation of action knowledge in multi-agent systems. Robot. Auton. Syst. 29(1), 41–50 (1999)CrossRefGoogle Scholar
- 9.Garland, A., Alterman, R.: Preparation of multi-agent knowledge for reuse. In: Proceedings of the Fall Symposium on Adaptation of Knowledge for Reuse, vol. 26, p. 33 (1995)Google Scholar
- 10.Gauci, J., Stanley, K.O.: Autonomous evolution of topographic regularities in artificial neural networks. Neural Comput. 22, 1860–1898 (2010)CrossRefMATHGoogle Scholar
- 11.Gauci, J., Stanley, K.: A case study on the critical role of geometric regularity in machine learning. In: AAAI, pp. 628–633 (2008)Google Scholar
- 12.Le Pape, C.: A combination of centralized and distributed methods for multi-agent planning and scheduling. In: Proceedings of the 1990 IEEE International Conference on Robotics and Automation, pp. 488–493. IEEE (1990)Google Scholar
- 13.Panait, L., Luke, S.: Cooperative multi-agent learning: the state of the art. Auton. Agent. Multi Agent Syst. 11(3), 387–434 (2005)CrossRefGoogle Scholar
- 14.Pugh, J.K., Goodell, S., Stanley, K.O.: Directional communication in evolved multiagent teams. Technical report, University of Central Florida, Orlando, FL (2013)Google Scholar
- 15.Sen, S., Weiss, G.: Learning in multiagent systems. In: Weiss, G. (ed.) Multiagent Systems: A Modern Approach to Distributed Artificial Intelligence, Chap. 6, pp. 259–298. The MIT Press, Cambridge (1999)Google Scholar
- 16.Stanley, K.O.: Compositional pattern producing networks: a novel abstraction of development. Genet. Program Evolvable Mach. 8, 131–162 (2007)CrossRefGoogle Scholar
- 17.Stanley, K.O., Bryant, B.D., Miikkulainen, R.: Real-time neuroevolution in the NERO video game. IEEE Trans. Evol. Comput. 9, 653–668 (2005)CrossRefGoogle Scholar
- 18.Stanley, K.O., D’Ambrosio, D.B., Gauci, J.: A hypercube-based encoding for evolving large-scale neural networks. Artif. Life 15, 185–212 (2009)CrossRefGoogle Scholar
- 19.Stanley, K.O., Miikkulainen, R.: A taxonomy for artificial embryogeny. Artif. Life 9, 93–130 (2003)CrossRefGoogle Scholar
- 20.Stanley, K.O., Miikkulainen, R.: Competitive coevolution through evolutionary complexification (2004)Google Scholar
- 21.Stone, P., Veloso, M.: Multiagent Systems: A Survey from a Machine Learning Perspective (2000)Google Scholar
- 22.Taylor, M.E., Whiteson, S., Stone, P.: Comparing evolutionary and temporal difference methods in a reinforcement learning domain. In: Proceedings of the 8th Annual Conference on Genetic and Evolutionary Computation GECCO 2006, p. 1321 (2006)Google Scholar
- 23.Turing, A.M.: The chemical basis of morphogenesis. Bull. Math. Biol. 52, 153–197 (1990)CrossRefGoogle Scholar
- 24.Verbancsics, P., Stanley, K.O.: Evolving static representations for task transfer. J. Mach. Learn. Res. 11, 1737–1769 (2010)MathSciNetMATHGoogle Scholar
- 25.Whiteson, S., Stone, P.: Evolutionary function approximation for reinforcement learning. J. Mach. Learn. Res. 7, 877–917 (2006)MathSciNetMATHGoogle Scholar
- 26.Whiteson, S., Whiteson, D.: Machine learning for event selection in high energy physics. Eng. Appl. Artif. Intell. 22, 1203–1217 (2009)CrossRefGoogle Scholar
- 27.Yao, X.: Evolving artificial neural networks (1999)Google Scholar
Copyright information
© Springer International Publishing Switzerland 2015