Abstract
The traditional command operation depend on the ground station is difficult to adapt to the highly dynamic and uncertain UAV air combat environment. Previous researches on UAV autonomous maneuver decision have limitations due to oversimplified assumptions, large and complex calculations, and lack of flexibility. Aiming at the air combat scenario of Red and Blue UAV, a three-dimensional UAV air combat model based on Markov Decision Process (MDP) is established. We trained Red UAV with Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithm to autonomously complete air combat missions. The performance is improved by scenario-transfer training and self-game training techniques. As the training scenario gradually transfers from simple to complex, the Red UAV continue to learn from previous experience steadily and improve the capabilities. The intelligence level is improved through self-game training. The simulation results present that the proposed maneuvering decision-making model and the training method can help the drone obtain effective decision-making strategies to get an advantage and defeat opponents in the confrontation.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Gupta, S.G., Ghonge, M., Jawandhiya, F.: Review of unmanned aircraft system (UAS). SSRN Electron. J. 2, 1646–1658 (2013)
Shin, H., Lee, J., Kim, H., Shim, D.H.: An autonomous aerial combat framework for two-on-two engagements based on basic fighter maneuvers. Aerosp. Sci. Technol. 72, 305–315 (2017)
Burgin, G.H., Owens, A.J.: An adaptive maneuvering logic computer program for the simulation of one-to-one air-to-air combat. Volume 2: Program description (1975)
Sutton, R.S., Barto, A.G.: Reinforcement learning: an introduction. IEEE Trans. Neural Netw. 9(5), 1054–1054 (1998)
Sutton, R.S., Mcallester, D., Singh, S., Mansour, Y.: Policy gradient methods for reinforcement learning with function approximation. In: Advances in Neural Information Processing Systems 12 (1999)
Mnih, V., Badia, A.P., Mirza, M., Graves, A., Kavukcuoglu, K.: Asynchronous methods for deep reinforcement learning (2016)
Lillicrap, T.P., et al.: Continuous control with deep reinforcement learning. Computer Science (2015)
Bernstein, D.S., Zilberstein, S., Immerman, N.: The complexity of decentralized control of Markov decision processes (2013)
Fujimoto, S., Hoof, H.V., Meger, D.: Addressing function approximation error in actor-critic methods (2018)
López, N.R., Bikowski, R.: Effectiveness of autonomous decision making for unmanned combat aerial vehicles in dogfight engagements. J. Guidance Control Dyn. 41, 1–7 (2018)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Jin, Q., Gao, X., Guo, Z., Hou, Z. (2022). Autonomous Maneuver Decision of UAV in Air Combat Based on Scenario-Transfer Deep Reinforcement Learning. In: Wu, M., Niu, Y., Gu, M., Cheng, J. (eds) Proceedings of 2021 International Conference on Autonomous Unmanned Systems (ICAUS 2021). ICAUS 2021. Lecture Notes in Electrical Engineering, vol 861. Springer, Singapore. https://doi.org/10.1007/978-981-16-9492-9_257
Download citation
DOI: https://doi.org/10.1007/978-981-16-9492-9_257
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-16-9491-2
Online ISBN: 978-981-16-9492-9
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)