Abstract
Agents have been employed to improve the performance of an Ant-Based Routing System on a communications network. The Agents use a Neural Net based Q-Learning approach to adapt their strategy according to conditions and learn autonomously. They are able to manipulate parameters that effect the behaviour of the Ant-System. The Ant-System is able to find the optimum routing configuration with static traffic conditions. However, under fast-changing dynamic conditions, such as congestion, the system is slow to react; due to the inertia built up by the best routes. The Agents reduce this inertia by changing the speed of response of the Ant-System. For an effective system, the Agents must cooperate – forming an implicit society across the network.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Brooks, R.: A Robust Layered Control System for a Mobile Agent. IEEE Journal of Robotics and Automation 2(1), 14–21 (1986)
Dorigo, M., Di Caro, G.: AntNet: A Mobile Agents Approach to Adaptive Routing., Technical Report IRIDIA 97-12 (1997)
Kapetanakis, S., Kudenko, D.: Improving on the Reinforcement Learning of Coordination in Cooperative Multi-Agent Systems. In: Proc. AAMAS-II, AISB 2002, pp. 89–94. Imperial College, London (2002)
Kapetanakis, S., Kudenko, D., Strens, M.: Learning to Coordinate using Commitment Sequences in Cooperative Multi-Agent Systems. In: Kudenko, D., Kazakov, D., Alonso, E. (eds.) AAMAS 2004. LNCS (LNAI), vol. 3394, pp. 106–118. Springer, Heidelberg (2005)
Schoonderwoerd, R., Holland, O., Bruten, J.: Ant-Like Agents for Load Balancing in Telecommunications Networks. In: Proc. 1st Int. Conf. on Autonomous Agents, Marina del Rey, pp. 209–216. ACM Press, New York (1997)
Schoonderwoerd, R., Holland, O., Bruten, J.: Ant-Based Load Balancing in Telecommunications Networks. Hewlett-Packard Laboratories, Bristol (1996)
Tesauro, G.: Programming backgammon using self-teaching neural nets. Artificial Intelligence 134(1-2), 181–199 (2002)
Tesauro, G.: Pricing in Agent Economies using Neural Networks and Multi-Agent Q-Lear ning. In: Proc. Workshop ABS-3 Learning About, From and With other Agents (August 1999)
Tumer, K., Wolpert, D.H.: Avoiding Braess’ Paradox through Collective Intelligence. AAAI/ IAAI (2000)
Vittori, K., Araújo, F.R.: Agent-Oriented Routing in Telecommunications Networks. IEICE Transactions on Communications E84-B(11), 3006–3013 (2001)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Legge, D. (2005). The Strategic Control of an Ant-Based Routing System Using Neural Net Q-Learning Agents. In: Kudenko, D., Kazakov, D., Alonso, E. (eds) Adaptive Agents and Multi-Agent Systems II. AAMAS AAMAS 2004 2003. Lecture Notes in Computer Science(), vol 3394. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-32274-0_10
Download citation
DOI: https://doi.org/10.1007/978-3-540-32274-0_10
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-25260-3
Online ISBN: 978-3-540-32274-0
eBook Packages: Computer ScienceComputer Science (R0)