Skip to main content
Log in

Deep reinforcement learning-based air combat maneuver decision-making: literature review, implementation tutorial and future direction

  • Published:
Artificial Intelligence Review Aims and scope Submit manuscript

Abstract

Nowadays, various innovative air combat paradigms that rely on unmanned aerial vehicles (UAVs), i.e., UAV swarm and UAV-manned aircraft cooperation, have received great attention worldwide. During the operation, UAVs are expected to perform agile and safe maneuvers according to the dynamic mission requirement and complicated battlefield environment. Deep reinforcement learning (DRL), which is suitable for sequential decision-making process, provides a powerful solution tool for air combat maneuver decision-making (ACMD), and hundreds of related research papers have been published in the last five years. However, as an emerging topic, there lacks a systematic review and tutorial. For this reason, this paper first provides a comprehensive literature review to help people grasp a whole picture of this field. It starts from the DRL itself and then extents to its application in ACMD. And special attentions are given to the design of reward function, which is the core of DRL-based ACMD. Then, a maneuver decision-making method based on one-to-one dogfight scenarios is proposed to enable UAV to win short-range air combat. The model establishment, program design, training methods and performance evaluation are described in detail. And the associated Python codes are available at gitee.com/wangyyhhh, thus enabling a quick-start for researchers to build their own ACMD applications by slight modifications. Finally, limitations of the considered model, as well as the possible future research direction for intelligent air combat, are also discussed.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16

Similar content being viewed by others

References

Download references

Acknowledgements

The authors are thankful to the financial support of the National Key Research and Development Plan (2021YFB3302501); the National Natural Science Foundation of China (12102077, 12161076, U2241263); the Fundamental Research Funds for the Central Universities (DUT22RC(3)010, DUT22LAB305, DUT22QN223, DUT22ZD211).

Author information

Authors and Affiliations

Authors

Contributions

XW: Conceptualization, Funding acquisition, Writing – original draft, Writing – review & editing, Supervision; YW: Writing – original draft, Visualization, Validation; XS: Project administration, Supervision; LW: Funding acquisition, Writing – review & editing; CL: Project administration; HP: Funding acquisition, Supervision; JL: Writing – review & editing.

Corresponding authors

Correspondence to Xinwei Wang or Xichao Su.

Ethics declarations

Competing interests

The authors declare no competing interests.

Code availability

The code is available at the URL gitee.com/wangyyhhh.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, X., Wang, Y., Su, X. et al. Deep reinforcement learning-based air combat maneuver decision-making: literature review, implementation tutorial and future direction. Artif Intell Rev 57, 1 (2024). https://doi.org/10.1007/s10462-023-10620-2

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10462-023-10620-2

Keywords

Navigation