Skip to main content

Trajectory Planning of UAV in Unknown Dynamic Environment with Deep Reinforcement Learning

  • Conference paper
  • First Online:
Proceedings of 2019 Chinese Intelligent Systems Conference (CISC 2019)

Part of the book series: Lecture Notes in Electrical Engineering ((LNEE,volume 593))

Included in the following conference series:

  • 976 Accesses

Abstract

Providing a collision-free, safe and efficient optimal trajectory for unmanned aerial vehicles (UAVs) in an unknown dynamic environment is one of the most important issues for researchers. In this paper, a trajectory planning approach for UAV in unknown dynamic environment based on deep reinforcement learning (DRL) is proposed. This study models trajectory planning of UAV as a discrete-time, discrete-action problem, and then proposes an improved deep Q network (IDQN) algorithm to solve it. The IDQN algorithm adds the track angle information of UAV to the reward function to speed up the learning process, furthermore, it also improves the action selection strategy and learning rate setting. Besides, in simulation, the paper considers the trajectory constraints of UAV in order to make the obtained trajectory have better practical availability. Simulation results demonstrate the effectiveness of the IDQN algorithm to implement UAV trajectory planning with constraints in unknown dynamic environments. Meanwhile, comparison with the classical DQN (CDQN) algorithm is conducted to further explore the advantage of the method.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 299.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Xu Z, Wei R, Zhang Q et al (2016) Obstacle avoidance algorithm for UAVs in unknown environment based on distributional perception and decision making. In: IEEE Chinese Guidance, Navigation and Control Conference (CGNCC). IEEE, pp 1072–1075

    Google Scholar 

  2. Zhao Y, Zheng Z, Liu Y (2018) Survey on computational-intelligence-based UAV path planning. Knowl-Based Syst 158:54–64

    Article  Google Scholar 

  3. Quan L (2012) Mobile robot path planning based on reinforcement learning in an unknown environment. Diss

    Google Scholar 

  4. Mnih V, Kavukcuoglu K, Silver D et al (2013) Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602

  5. Mnih V, Kavukcuoglu K, Silver D et al (2015) Human-level control through deep reinforcement learning. Nature 518(7540):529

    Article  Google Scholar 

  6. Lillicrap TP, Hunt JJ, Pritzel A et al (2015) Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971

  7. Heess N, Sriram S, Lemmon J et al (2017) Emergence of locomotion behaviours in rich environments. arXiv preprint arXiv:1707.02286

  8. Mnih V, Badia AP, Mirza M et al (2016) Asynchronous methods for deep reinforcement learn-ing. In: International conference on machine learning, pp 1928–1937

    Google Scholar 

  9. Jaderberg M, Mnih V, Czarnecki WM et al (2016) Reinforcement learning with unsupervised auxiliary tasks. arXiv preprint arXiv:1611.05397

  10. Polvara R, Patacchiola M, Sharma S et al (2017) Autonomous quadrotor landing using deep reinforcement learning. arXiv preprint arXiv:1709.03339

  11. Huang H, Yang Y, Gui G et al (2019) Deep reinforcement learning for UAV Navigation through Massive MIMO. arXiv preprint arXiv:1901.10832

  12. Rodriguez-Ramos A, Sampedro C, Bavle H et al (2019) A deep reinforcement learning strategy for UAV autonomous landing on a moving platform. J Intell Robot Syst 93(1–2):351–366

    Article  Google Scholar 

  13. Zeng J, Qin L, Hu Y et al (2019) Integrating a path planner and an adaptive motion controller for navigation in dynamic environments. Appl Sci 9(7):1384

    Article  Google Scholar 

  14. Wang X, Jiang Z, Xie R, Chen H (2012) Engine constraints in UAV route planning (Doctoral Dissertation)

    Google Scholar 

  15. Ng AY, Jordan MI (2003) Shaping and policy search in reinforcement learning. University of California, Berkeley

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jia Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Wang, J., Wang, W., Wu, Q. (2020). Trajectory Planning of UAV in Unknown Dynamic Environment with Deep Reinforcement Learning. In: Jia, Y., Du, J., Zhang, W. (eds) Proceedings of 2019 Chinese Intelligent Systems Conference. CISC 2019. Lecture Notes in Electrical Engineering, vol 593. Springer, Singapore. https://doi.org/10.1007/978-981-32-9686-2_54

Download citation

Publish with us

Policies and ethics