Skip to main content
Log in

Local Planning Strategy Based on Deep Reinforcement Learning Over Estimation Suppression

  • Engine and Emissions, Fuels and Lubricants
  • Published:
International Journal of Automotive Technology Aims and scope Submit manuscript

Abstract

Local planning is a critical and difficult task for intelligent vehicles in dynamic transportation environments. In this paper, a new method Suppress Q Deep Q Network (SQDQN) combining traditional deep reinforcement learning Deep Q Network (DQN) with information entropy is proposed for local planning in automatic driving. In the proposed approach, local planning strategy in complex traffic environment established by the actor–critic network based on DQN, the method adopts the way of execution action-evaluation action-update network to explore the optimal local planning strategy. Proposed strategy does not rely on accurate modeling of the scene, so it is suitable for complex and changeable traffic scenes. At the same time, evaluate the update process and determine the update range by using information entropy to solve a common problem in the network that over expectation of actions damage the performance of strategies. Use this approach to improve strategic performance. The trained local planning strategy is evaluated in three simulation scenarios: overtaking, following, driving in hazardous situations. The results illustrate the advantages of the proposed SQDQN method in solving local planning problem.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

Data availability

The data used in this study are available and willing to be provided to others for use upon reasonable request.

References

  • Arulkumaran, K., Deisenroth, M. P., Brundage, M., & Bharath, A. A. (2017). Deep reinforcement learning: A brief survey. IEEE Signal Processing Magazine, 34(6), 26–38.

    Article  Google Scholar 

  • Badrinarayanan, V., Kendall, A., & Cipolla, R. (2017). SegNet: A deep convolutional encoder–decoder architecture for image segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39(12), 2481–2495.

    Article  Google Scholar 

  • Chu, K., Lee, M., & Sunwoo, M. (2012). Local path planning for off-road autonomous driving with avoidance of static obstacles. IEEE Transactions on Intelligent Transportation Systems, 13(4), 1599–1616.

    Article  Google Scholar 

  • Gao, H. B., Zhu, J. P., Zhang, T., Xie, G. T., Kan, Z., Hao, Z. Y., & Liu, K. (2022). Situational assessment for intelligent vehicles based on stochastic model and Gaussian distributions in typical traffic scenarios. IEEE Transactions on Systems, Man, and Cybernetics: System, 52(3), 1426–1436.

    Article  Google Scholar 

  • Gao, S. H., Huang, S. N., Xiang, C., & Lee, T. H. (2020). A review of optimal motion planning for unmanned vehicles. Journal of Marine Science and Technology-Taiwan, 28(25), 321–330.

    Google Scholar 

  • Hassaballah, M., Kenk, M. A., Muhammad, K., & Minaee, S. (2021). Vehicle detection and tracking in adverse weather using a deep learning framework. IEEE Transactions on Intelligent Transportation Systems, 22(7), 4230–4242.

    Article  Google Scholar 

  • Huang, Y. J., Ding, H. T., Zhang, Y. B., Wang, H., Cao, D. P., Xu, N., & Hu, C. (2020). A motion planning and tracking framework for autonomous vehicles based on artificial potential field elaborated resistance network approach. IEEE Transactions on Industrial Electronics, 67(2), 1376–1386.

    Article  Google Scholar 

  • Jeong, Y., & Yi, K. (2021). Target vehicle motion prediction-based motion planning framework for autonomous driving in uncontrolled intersections. IEEE Transactions on Intelligent Transportation Systems, 22(1), 168–177.

    Article  Google Scholar 

  • Kong, D. W., List, G. F., Guo, X. C., & Wu, D. X. (2018). Modeling vehicle car-following behavior in congested traffic conditions based on different vehicle combinations. Transportation Letters-the International Journal of Transportation Research, 10(5), 280–293.

    Google Scholar 

  • Li, L. J., Gan, Z. Y., Xu, Qu., & Ran, B. (2020). Risk perception and the warning strategy based on safety potential field theory. Accident Analysis & Prevention, 148, 1–17.

    Article  Google Scholar 

  • Li, X., Sun, Z., Dongpu, C., Liu, D., & He, H. (2017a). Development of a new integrated local trajectory planning and tracking control framework for autonomous ground vehicles. Mechanical Systems and Signal Processing, 87, 118–137.

    Article  Google Scholar 

  • Li, Y., Wang, Jq., & Wu, J. (2017b). Model calibration concerning risk coefficients of driving safety field model. Journal of Central South University, 24(1), 1494–1502.

    Article  Google Scholar 

  • Lopez-Martin, M., Carro, B., & Sanchez-Esguevillas, A. (2020). Application of deep reinforcement learning to intrusion detection for supervised problems. Expert Systems with Applications, 141, 1–15.

    Article  Google Scholar 

  • Morales, E. F., Murrieta-Cid, R., Becerra, I., & Esquivel-Basaldua, M. A. (2021). A survey on deep learning and deep reinforcement learning in robotics with a tutorial on deep reinforcement learning. Intelligent Service Robotics, 14(5), 773–805.

    Article  Google Scholar 

  • Rafiei, A., Fasakhodi, A. O., & Hajati, F. (2022). Pedestrian collision avoidance using deep reinforcement learning. International Journal of Automotive Technology, 23, 613–622.

    Article  Google Scholar 

  • Sajjad, M., Irfan, M., Muhammad, K., Del Ser, J., Sanchez-Medina, J., Andreev, S., Ding, W. P., & Lee, J. W. (2021). An efficient and scalable simulation model for autonomous vehicles with economical hardware. IEEE Transactions on Intelligent Transportation Systems, 22(3), 1718–1732.

    Article  Google Scholar 

  • Sakhare, K. V., Tewari, T., & Vyas, V. (2020). Review of vehicle detection systems in advanced driver assistant systems. Archives of Computational Methods in Engineering, 27(2), 591–610.

    Article  Google Scholar 

  • van Hasselt, H. (2010). Double Q-learning[C]. In 23rd Advances in Neural Information Processing Systems (NeurIPS). Vancouver, British Columbia: MIT Press, pp. 2613–2621.

  • Wang, H., Huang, Y. A., Khajepour, Y., Zhang, Y. R., & Cao, D. (2019). Crash mitigation in motion planning for autonomous vehicles. IEEE Transactions on Intelligent Transportation Systems, 20(9), 3313–3323.

    Article  Google Scholar 

  • Wang, H., Wang, Q., Chen, W., Zhao, L., & Tan, D. (2021). Path tracking based on model predictive control with variable predictive horizon. Transactions of the Institute of Measurement and Control, 43(12), 2676–2688.

    Article  Google Scholar 

  • Wang, Liu, X. J., Qiu, T., Mu, C., Chen, C., & Zhou, P. (2020). A real-time collision prediction mechanism with deep learning for intelligent transportaton system. IEEE Transactions on Vehicular Technology, 69(9), 9497–9508.

    Article  Google Scholar 

  • Watkins, C. J. C. H. (1989). Learning from delayed rewards[D]. University of Cambridge.

    Google Scholar 

  • Wen, S., Zhao, Y., Yuan, X., Wang, Z., Zhang, D., & Manfredi, L. (2020). Path planning for active SLAM based on deep reinforcement learning under unknown environments. Intelligent Service Robotics, 13, 263–272.

    Article  Google Scholar 

  • Yoo, J. M., Jeong, Y., & Yi, K. (2021). Virtual target-based longitudinal motion planning of autonomous vehicles at urban intersections: determining control inputs of acceleration with human driving characteristic-based constraints. IEEE Vehicular Technology Magazine, 16(3), 38–46.

    Article  Google Scholar 

  • Zhang, X., Ma, H., Luo, X., & Yuan, J. (2022). LIDAR: Learning from imperfect demonstrations with advantage rectification. Frontiers of Computer Science, 16(1), 1–9.

    Article  Google Scholar 

  • Zhao, Z. Y., Wang, Q., & Li, X. L. (2020). Deep reinforcement learning based lane detection and localization. Neurocomputing, 413, 328–338.

    Article  Google Scholar 

  • Zuo, Z. Q., Yang, X., Li, Z., Wang, Y. J., Han, Q. N., Wang, L., & Luo, X. Y. (2021). MPC-based cooperative control strategy of path planning and trajectory tracking for intelligent vehicles. IEEE Transactions on Intelligent Vehicles, 6(3), 513–522.

    Article  Google Scholar 

Download references

Acknowledgements

The authors are very grateful to the China government by the support of this work through National Natural Science Foundation Program (52372407), Jilin Provincial Science and Technology Development Plan Project (20230402064GH) and the Natural Science Foundation of Jilin Province (20220101236JC).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Changsheng Zhu.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Han, L., Wang, Y., Chi, R. et al. Local Planning Strategy Based on Deep Reinforcement Learning Over Estimation Suppression. Int.J Automot. Technol. (2024). https://doi.org/10.1007/s12239-024-00076-w

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s12239-024-00076-w

Keywords

Navigation