Skip to main content

Deep Reinforcement Learning-Based Intelligent Decision-Making for Orbital Game of Satellite Swarm

  • Conference paper
  • First Online:
Computational and Experimental Simulations in Engineering (ICCES 2023)

Part of the book series: Mechanisms and Machine Science ((Mechan. Machine Science,volume 145))

  • 608 Accesses

Abstract

Recent years have witnessed the rapid development of aerospace science and technology, and the orbital game technology has shown great potential value in the field of failed satellite maintenance, debris removal, etc. In this case, orbital game is often characterized by nonlinear dynamic model, unknown state information, high randomness, but the existing approaches to deal with game problem are difficult to be applied. The analytical method based on game theory is only applicable to simple scenarios, and it is challenging to find the optimal strategy for such complex scenarios as satellite swarm game. It should be noted that deep reinforcement learning has some research basis in the cooperative decision-making and control of multi-agents. In view of its powerful perception and decision ability, this paper applies deep reinforcement learning to solve the orbital game problem of satellite swarm. Firstly, the game scenario is modeled, where typical constraints, e.g., minimum time, optimal fuel, and collision avoidance, are taken into consideration in the game process, and then the multi-agent reinforcement learning algorithm is developed to solve the optimal maneuver strategy. The algorithm is based on the Actor-Critic architecture and uses a centralized training and decentralized execution approach to solve the optimal joint maneuver strategy. For different task scenarios, the action space, state observation space, and reward space are designed to introduce more rewards that match the specific game tasks to make the algorithm converge quickly, so that the satellite swarm emerges and executes better intelligent strategies to complete the corresponding game task.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 219.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
USD 279.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Zhao, L.R., Dang, Z.H., Zhang, Y.L.: Orbital game: concepts, principles and methods. J. Command Control 7(3), 215–224 (2021)

    Google Scholar 

  2. Zhou, J.: Manned space rendezvous and docking technology. Manned Space 17(2), 1–8 (2011)

    Google Scholar 

  3. Yuan, L.: Spacecraft intelligent autonomous control technology toward uncertain environment. J. Astronaut. 42(7), 839–849 (2021)

    Google Scholar 

  4. Sun, S., Zhang, Q., Loxton, R., et al.: Numerical solution of a pursuit-evasion differential game involving two spacecraft in low earth orbit. J. Ind. Manag. Optim. (JIMO) 11(4), 1127–1147 (2015)

    Article  MathSciNet  MATH  Google Scholar 

  5. Li, Z., Zhu, H., Yang, Z., et al.: Saddle point of orbital pursuit-evasion game under J2-perturbed dynamics. J. Guid. Control Dyn. 43(9), 1733–1739 (2020)

    Article  Google Scholar 

  6. Pontani, M., Conway, B.A.: Numerical solution of the three-dimensional orbital pursuit-evasion game. J. Guid. Control Dyn. 2(32), 474–487 (2009)

    Article  Google Scholar 

  7. Stupik, J., Pontani, M., Conway, B.: Optimal pursuit/evasion spacecraft trajectories in the hill reference frame. In: AIAA/AAS Astrodynamics Specialist Conference, p. 4882 (2012)

    Google Scholar 

  8. Hafer, W.T., Reed, H.L., Turner, J.D., et al.: Sensitivity methods applied to orbital pursuit evasion. J. Guid. Control Dyn. 38(6), 1118–1126 (2015)

    Article  Google Scholar 

  9. Liu, Y., Ye, D., Hao, Y.: Distributed online mission planning for multi-player space pursuit and evasion. Chin. J. Aeronaut. 29(6), 1709–1720 (2016)

    Article  Google Scholar 

  10. Harmon, M.E., Baird, L.C., Klopf, A.H.: Reinforcement learning applied to a differential game. Adapt. Behav. 4(1), 3–28 (1995)

    Article  Google Scholar 

  11. Liu, B.Y., Ye, X.B., Gao, Y., et al.: Forward-looking imaginative planning framework combined with prioritized replay double DQN. In: International Conference Control, Automation and Robotics, pp. 336–341 (2019)

    Google Scholar 

  12. Xu, D., Hui, Z., Liu, Y.Q., et al.: Morphing control of a new bionic morphing UAV with deep reinforcement learning. Aerosp. Sci. Technol. 92, 232–243 (2019)

    Article  Google Scholar 

  13. Zhu, Y., Mottaghi, R., Kolve, E., et al.: Target-driven visual navigation in indoor scenes using deep reinforcement learning. In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 3357–3364. IEEE (2017)

    Google Scholar 

  14. Liu, B., Ye, X., Gao, Y., et al.: Strategy solution of non-cooperative target pursuit-evasion game based on branching deep reinforcement learning. Acta Aeronaut. Astronaut. Sinica 41(10), 348–358 (2020)

    Google Scholar 

  15. Wang, C., Ye, D., Sun, Z., et al.: Adaptive game strategy of spacecraft terminal interception. J. Astronaut. 41(3), 309–318 (2020)

    Google Scholar 

  16. Lowe, R., Wu, Y., Tamar, A., et al.: Multi-agent actor-critic for mixed cooperative-competitive environments. Advances in Neural Information Processing Systems, pp. 6379–6390 (2017)

    Google Scholar 

  17. Mnih, V., Kavukcuoglu, K., Silver, D., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529–533 (2015)

    Google Scholar 

  18. Lillicrap, T., Hunt, J., Pritzel, A., et al.: Continuous control with deep reinforcement learning. Comput. Sci. (2015)

    Google Scholar 

  19. Zavoli, A., Federici, L.: Reinforcement learning for robust trajectory design of interplanetary missions. J. Guid. Control Dyn. 44(8), 1440–1453 (2021)

    Article  Google Scholar 

  20. Hovell, K., Ulrich, S.: Deep reinforcement learning for spacecraft proximity operations guidance. J. Spacecr. Rocket. 58(2), 254–264 (2021)

    Article  Google Scholar 

  21. Shirobokov, M., Trofimov, S., Ovchinnikov, M.: Survey of machine learning techniques in spacecraft control design. Acta Astronaut. 186, 87–97 (2021)

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported in part by the National Natural Science Foundation of China under Grants U21B6001, 11972026 and U2013206, in part by Science and Technology on Space Intelligent Control Laboratory under Grant 2021-JCJQ-LB-010-07, and in part by Key Research and Development Program of Shaanxi under Grant 2023-YBGY-384.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Chuang Liu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yu, W., Yue, X., Huang, P., Liu, C. (2024). Deep Reinforcement Learning-Based Intelligent Decision-Making for Orbital Game of Satellite Swarm. In: Li, S. (eds) Computational and Experimental Simulations in Engineering. ICCES 2023. Mechanisms and Machine Science, vol 145. Springer, Cham. https://doi.org/10.1007/978-3-031-42987-3_61

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-42987-3_61

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-42986-6

  • Online ISBN: 978-3-031-42987-3

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics