Abstract
This paper studies a path planning method for multiple robots in unknown environment. Multiple robots adopt the leader-following formation method. For the Q-learning algorithm used by the leader robot, the Q-table is initialized by prior information of environment and the idea of filling concave obstacles is proposed. Then the strategy of choosing actions is improved by simulated annealing algorithm, which changes the greedy factor in real time according to the Q-learning. The follower robot uses an improved gravitational potential field method to follow the leader robot. The simulation results show that the improved algorithm is effective and multiple robots can plan an optimum path to reach the destination with this method.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Igarashi, H.: Path planning of a mobile robot by optimization and reinforcement learning. Artif. Life Rob. 6, 59–65 (2002). https://doi.org/10.1007/BF02481210
Wang, D., Deng, H., Pan, Z.: MRCDRL: multi-robot coordination with deep reinforcement learning. Neurocomputing 406, 68–76 (2020). https://doi.org/10.1016/j.neucom.2020.04.028
Zhang, T., Zhu, Y., Song, J.: Real-time motion planning for mobile robots by means of artificial potential field method in unknown environment. Ind. Rob. 37, 384–400 (2013). https://doi.org/10.1108/01439911011044840
Ye, D., Zhu, T., Cheng, Z., et al.: Differential advising in multiagent reinforcement learning. IEEE Trans. Cybern. 99, 1–14 (2020). https://doi.org/10.1109/TCYB.2020.3034424
Dong, P.F., Zhang, Z.A., Mei, X.H., et al.: Reinforcement learning path planning algorithm with the introduction of potential field and trap search. Comput. Eng. Appl. 54, 129–134 (2018)
Xu, X.S., Yuan, J.: Path planning method of mobile robot based on improved reinforcement learning. J. Chin. Inertial Technol. 27, 314–320 (2019)
Sadhu, A.K., Konar, A.: Improving the speed of convergence of multi-agent Q-learning for cooperative task-planning by a robot-team. Rob. Auton. Syst. 92, 66–80 (2017). https://doi.org/10.1016/j.robot.2017.03.003
Cruz, D.L., Yu, W.: Path planning of multi-agent systems in unknown environment with neural kernel smoothing and reinforcement learning. Neurocomputing 233, 34–42 (2017). https://doi.org/10.1016/j.neucom.2016.08.108
Tian, S., Li, Y., Kang, Y., et al.: Multi-robot path planning in wireless sensor networks based on jump mechanism PSO and safety gap obstacle avoidance. Future Gener. Comput. Syst. 118, 37–47 (2020). https://doi.org/10.1016/j.future.2020.12.012
Acknowledgements
The work is supported by the National Natural Science Foundation of China (61673200, 61903172), the Major Basic Research Project of Natural Science Foundation of Shandong Province of China (ZR2018ZC0438) and the Key Research and Development Program of Yantai City of China (2019XDHZ085).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Fan, Z., Yang, H., Han, Y., Ning, X. (2022). A Path Planning Method for Multi-robot Formation Based on Improved Q-Learning. In: Jia, Y., Zhang, W., Fu, Y., Yu, Z., Zheng, S. (eds) Proceedings of 2021 Chinese Intelligent Systems Conference. Lecture Notes in Electrical Engineering, vol 805. Springer, Singapore. https://doi.org/10.1007/978-981-16-6320-8_87
Download citation
DOI: https://doi.org/10.1007/978-981-16-6320-8_87
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-16-6319-2
Online ISBN: 978-981-16-6320-8
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)