Abstract
Learning a Nash equilibrium of a game is challengeable, and the issue of learning all the Nash equilibria seems intractable. This paper investigates the effectiveness of a momentum-based approach to learning the Nash equilibria of a game. Experimental results show the proposed algorithm can learn a Nash equilibrium in each learning iteration for a normal form strategic game. By employing a deflection technique, it can learn almost all the existing Nash equilibria of a game.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Vrieze, O.J., Tij, S.H.: Fictitious play applied to sequence of games and discounted stochastic games. International Journal of Game Theory 11(2), 71–85 (1982)
Claus, C., Boutilier, C.: The dynamics of reinforcement learning in cooperative multiagent systems. In: proceedings of 15th National Conference on Artificial Intelligence (1998)
Verbeeck, K., Nowe, A., Lenaerts, T., Parent, J.: Learning to reach the pareto optimal Nash Equilibrium as a Team. In: McKay, B., Slaney, J.K. (eds.) Canadian AI 2002. LNCS (LNAI), vol. 2557, Springer, Heidelberg (2002)
Littman, M.L.: Markov games as a framework for multiagent reinforcement learning. In: Proc: 11th international conference on machine learning, New Brunswick, NJ, pp. 157–163. Morgan Kaufmann, San Mateo (1994)
Hu, J., Wellman, M.P.: Nash Q-Learning for General-Sum Stochastic Games. Journal of Machine Learning research 1, 1–30 (2003)
Bowling, M., Veloso, M.: Multiagent learning using a variable learning rate. Artificial Intelligence 136, 215–250 (2002)
Singh, S., Kearns, M., Mansouv, Y.: Nash convergence of gradient dynamics in general-sum games. In: proceedings of the 17th conference on uncertainty in artificial intelligence, pp. 541–548 (2000)
Zhang, H., Huang, S.: Convergent Gradient Ascent with Momentum in General-Sum Games. Neurocomputing 61, 445–449 (2004)
Mckevey, R.D.: A Liapunov function for Nash equilibria. Technical Report, California institute of Technical Report, California Institute of Technology (1991)
Mckelvey, R.D., Mclennan, A., Turocy, T.: Gambit Command language, California Institute of Technology (2000)
Hansan, N., Stermer, A.: Completely derandomized self-adaption in evolution strategies. Evolutionary Computation 9(2), 159–195 (2001)
Kennedy, J., Eberhart, R.C.: Swarm intelligence. Morgan Kaufmann Publishers, Los Altos (2001)
Pavlids, N.G., Parsopoulos, K.E., Vrahatis, M.N.: Computing Nash equilibria through computational intelligence methods 175, 113–136 (2005)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zhang, H., Liu, P. (2006). A Momentum-Based Approach to Learning Nash Equilibria. In: Shi, ZZ., Sadananda, R. (eds) Agent Computing and Multi-Agent Systems. PRIMA 2006. Lecture Notes in Computer Science(), vol 4088. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11802372_54
Download citation
DOI: https://doi.org/10.1007/11802372_54
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-36707-9
Online ISBN: 978-3-540-36860-1
eBook Packages: Computer ScienceComputer Science (R0)