Abstract
This paper relieves the ‘curse of dimensionality’ problem, which becomes intractable when scaling reinforcement learning to multi-agent systems. This problem is aggravated exponentially as the number of agents increases, resulting in large memory requirement and slowness in learning speed. For cooperative systems which widely exist in multi-agent systems, this paper proposes a new multi-agent Q-learning algorithm based on decomposing the joint state and joint action learning into two learning processes, which are learning individual action and the maximum value of the joint state approximately. The latter process considers others’ actions to insure that the joint action is optimal and supports the updating of the former one. The simulation results illustrate that the proposed algorithm can learn the optimal joint behavior with smaller memory and faster learning speed compared with friend-Q learning and independent learning.
Similar content being viewed by others
References
G. Weiss. Multiagent Systems: A Modern Approach to Distributed Artificial Intelligence. Cambridge: MIT Press, 1999.
N. Vlassis. A concise introduction to multiagent systems and distributed artificial intelligence. Synthesis Lectures on Artificial Intelligence and Machine Learning, 2007, 1(1): 1–71.
M. Wu, W. Cao, J. Peng, et al. Balanced reactive-deliberative architecture for multi-agent system for simulation league of RoboCup. International Journal of Control, Automation and Systems, 2009, 7(6): 945–955.
K. Tumer, A. Agogino. Improving air traffic management with a learning multiagent system. IEEE Intelligent Systems, 2009, 24(1):18–21.
S. Proper, P. Tadepalli. Solving multiagent assignment Markov decision processes. Proceedings of the 8th International Joint Conference on Autonomous Agents and Multiagent Systems. Richland: IFAAMAS, 2009: 681–688.
J. R. Kok, M. T. J. Spaan, N. Vlassis. Non-communicative multi-robot coordination in dynamics environments. Robotics and Autonomous Systems, 2005, 50(2/3): 99–114.
M. L. Littman. Friend-or-Foe Q-learning in general-sum games. Proceedings of the 18th International Conference on Machine Learning. Williamstown: Morgan Kaufmann Press, 2001: 322–328.
X. Wang, T. Sandholm. Reinforcement learning to play an optimal Nash equilibrium in team Markov games. Proceedings of the Advances Neural Information Processing Systems. Cambridge: MIT Press, 2002: 1571–1578.
R. I. Brafman, M. Tennenholtz. R-Max-a general polynomial time algorithm for near-optimal reinforcement learning. Journal of Machine Learning Research, 2002, 3(2): 213–231.
L. Busoniu, R. Babuska, B. De Schutter. A comprehensive survey of multi-agent reinforcement learning. IEEE Transactions on Systems, Man, and Cybernetics - Part C: Applications and Reviews, 2008, 38(2): 156–172.
N. Mehta, S. Natarajan, P. Tadepalli, et al. Transfer in variable-reward hierarchical reinforcement learning. Machine Learning, 2008, 73(3):289–312.
J. R. Kok, N. Vlassis. Collaborative multiagent reinforcement learning by payoff propagation. Journal of Machine Learning Research, 2006, 7: 1789–1828.
S. Kapetanakis, D. Kudenko. Reinforcement learning of coordination in cooperative multi-agent systems. Proceedings of the 18th National Conference on Artificial Intelligence. Washington: IEEE Computer Society, 2002: 326–331.
C. Claus, C. Boutilier. The dynamics of reinforcement learning in cooperative multiagent systems. Proceedings of the 15th National Conference on Artificial Intelligence. Madison: AAAI Press, 1998:746–752. 746–752.
C. J. C. H. Watkins, P. Dayan. Q-learning. Machine Learning, 1992, 8(3/4): 279–292.
C. S. Szepesvari, M. L. Littman. A unified analysis of value-function-based reinforcement-learning algorithms. Neural Computation, 1999, 11(8): 2017–2059.
R. S. Sutton. Learning to predict by the method of temporal differences. Machine Learning, 1988, 3(1): 9–44.
A. Bab, R. I. Brafman. Multi-agent reinforcement learning in common interest and fixed sum stochastic games: an experimental study. Journal of Machine Learning Research, 2008, 9: 2635–2675.
Author information
Authors and Affiliations
Corresponding author
Additional information
This work was supported by National Nature Science Foundation of China (Nos. 61074058, 60874042), the Chinese Postdoctoral Science Foundation (No. 200902483), the Specialized Research Fund for the Doctoral Program of Higher Education of China (No. 20090162120068), and the Central South University Innovation Project (No. 2011ssxt221).
Xin CHEN received his B.S. degree in Industrial Automation, and M.S. degree in Control Theory and Control Engineering from Central South University, Changsha, China in 1999 and 2002, respectively. He received his Ph.D. in Electromechanical Engineering from the University of Macau, China in 2007. He is currently an associate professor at Central South University. His research interests include multi-agent system, robotics and intelligent control.
Gang CHEN received his B.S. degree in Automation from Central South University, Changsha, China in 2009. He is currently working toward the M.S. degree in Control Theory and Control Engineering at School of Information and Science, Central South University. His research interests include multi-agent system, reinforcement learning.
Weihua CAO received his B.S., M.S., and Ph.D. degrees in Engineering from Central South University, Changsha, China, in 1994, 1997, and 2007, respectively. Since May 1997, he has been a faculty member with Central South University, where he is currently a professor of automatic control engineering at the School of Information Science and Engineering. He was a visiting student with the Department of Engineering, Kanazawa University, Japan, from 1996 to 1997, and a Visiting Scholar with the Department of Electrical and Computer Engineering, University of Alberta, Edmonton, AB, Canada, during the 2007–2008 academic year. His research interests include multi-agent system, intelligent control and process control.
Min WU received his B.S. and M.S. degrees in Engineering from Central South University, Changsha, China, in 1983 and 1986, respectively, and Ph.D. degree in Engineering from Tokyo Institute of Technology, Tokyo, Japan, in 1999. Since July 1986, he has been a faculty member with Central South University, where he is currently a professor of Automatic Control Engineering at the School of Information Science and Engineering. He was a visiting scholar with the Department of Electrical Engineering, Tohoku University, Sendai, Japan, from 1989 to 1990, and a visiting research scholar with the Department of Control and Systems Engineering, Tokyo Institute of Technology, from 1996 to 1999. His current research interests include robust control and its applications, process control, and intelligent control.
Dr. Wu is a member of the Nonferrous Metals Society of China and the Chinese Association of Automation. He received the IFAC Control Engineering Practice Prize Paper Award in 1999 (together with M. Nakano and J. She). Dr. Wu is a senior member of the IEEE.
Rights and permissions
About this article
Cite this article
Chen, X., Chen, G., Cao, W. et al. Cooperative learning with joint state value approximation for multi-agent systems. J. Control Theory Appl. 11, 149–155 (2013). https://doi.org/10.1007/s11768-013-1141-z
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11768-013-1141-z