Skip to main content
Log in

Improving environmental awareness for autonomous vehicles

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Autonomous vehicles (AVs) have multiple tasks with different priorities and safety levels where classic supervised learning techniques are no longer applicable. Thus, reinforcement learning (RL) algorithms become increasingly appropriate for this domain as the RL algorithms can act on complex problems and adapt their responses in the face of unforeseen situations and environments. The RL agent aims to perform the action that guarantees the optimal reward with the best score. The problem with this approach is if the agent finds a possible optimal action with a reasonable premium and gets stuck in this mediocre strategy, which at the same time is neither the best nor the worst solution. Therefore, the agent avoids performing a more extensive exploration to find new paths and learn alternatives to generate a higher reward. To alleviate this problem, we research the behavior of two types of noise in AVs training. We analyze the results and point out the noise method that most stimulates exploration. A vast exploration of the environment is highly relevant to AVs because they know more about the environment and learn alternative ways of acting in the face of uncertainties. With that, AVs can expect more reliable actions in front of sudden changes in the environment. According to our experiments’ results in a simulator, we can see that noise allows the autonomous vehicle to improve its exploration and increase the reward.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  1. Liu X, Xu H, Liao W, Yu W (2019) Institute of Electrical and Electronics Engineers Inc., pp 318–327. https://doi.org/10.1109/ICII.2019.00063

  2. Lillicrap TP, Hunt JJ, Pritzel A, Heess N, Erez T, Tassa Y, Silver D, Wierstra D (2015) arXiv:1509.02971

  3. Kuderer M, Gulati S, Burgard W (2015) Institute of Electrical and Electronics Engineers Inc., pp 2641–2646. https://doi.org/10.1109/ICRA.2015.7139555

  4. Lindemann P, Lee T-Y, Rigoll G (2018) . Multimodal Technol Interact 2:4. https://doi.org/10.3390/mti2040071, https://www.mdpi.com/2414-4088/2/4/71

    Article  Google Scholar 

  5. Ziebart BD (2010) Modeling purposeful adaptive behavior with the principle of maximum causal entropy

  6. Zhu H, Yuen K-V, Mihaylova L, Leung H (2017) . IEEE Trans Intell Transp Syst 18 (10):2584–2601. https://doi.org/10.1109/TITS.2017.2658662

    Article  Google Scholar 

  7. Ahmed Z, Roux NL, Norouzi M, Schuurmans D (2018) arXiv:1811.11214

  8. Leite A, Candadai M, Izquierdo EJ (2020) . Artif Life Conf Proc 32:441–449. https://doi.org/10.1162/isal_a_00338

    Google Scholar 

  9. Nair A, McGrew B, Andrychowicz M, Zaremba W, Abbeel P (2018) Institute of Electrical and Electronics Engineers Inc., pp 6292–6299. https://doi.org/10.1109/ICRA.2018.8463162

  10. Stekolshchik R (2020) Noise, overestimation and exploration in deep reinforcement learning

  11. Nauta J, Khaluf Y, Simoens P Using the ornstein-uhlenbeck process for random exploration

  12. Bougie1 N, Ichise R (2021) Fast and slow curiosity for high-level exploration in reinforcement learning. https://doi.org/10.1007/s10489-020-01849-3

  13. Pong VH, Dalal M, Lin S, Nair A, Bahl S, Levine S (2019) arXiv:1903.03698

  14. Hong Z-W, Shann T-Y, Su S-Y, Chang Y-H, Lee C-Y (2018) arXiv:1802.04564

  15. Iyengar K, Dwyer G, Stoyanov D (2020) . Int J CARS 15:1157–1165. https://doi.org/10.1007/s11548-020-02194-z

    Article  Google Scholar 

  16. Brockman G, Cheung V, Pettersson L, Schneider J, Schulman J, Tang J, Zaremba W (2016) Openai gym

  17. Hafner D, Lillicrap T, Norouzi M, Ba J (2021) Mastering atari with discrete world models

  18. Schulman J, Wolski F, Dhariwal P, Radford A, Klimov O (2017) Proximal policy optimization algorithms

  19. Stanton C, Clune J (2018) Deep curiosity search: Intra-life exploration can improve performance on challenging deep reinforcement learning problems

  20. Ferret J, Pietquin O, Geist M (2021) Self-imitation advantage learning. www.ifaamas.org

  21. Mnih V, Badia AP, Mirza M, Graves A, Lillicrap TP, Harley T, Silver D, Kavukcuoglu K (2016) Asynchronous methods for deep reinforcement learning

  22. Bellemare MG, Srinivasan S, Ostrovski G, Schaul T, Saxton D, Munos R (2016) Unifying count-based exploration and intrinsic motivation

  23. Plappert M, Houthooft R, Dhariwal P, Sidor S, Chen RY, Chen X, Asfour T, Abbeel P, Andrychowicz M (2017) arXiv:1706.01905

  24. Ghafoorianfar N, Roopaei M (2020) 2020 10th Annual Computing and Communication Workshop and Conference (CCWC), pp 0444–0448

  25. Rasmussen CE (2004). In: Bousquet O, von Luxburg U, Rätsch G (eds) Gaussian processes in machine learning. Springer, Berlin

  26. Uhlenbeck GE, Ornstein LS (1930) . Phys Rev 36:823–841. https://doi.org/10.1103/PhysRev.36.823

    Article  Google Scholar 

  27. Bartoszek K, Glémin S, Kaj I, Lascoux M (2017) . J Theor Biol 429:35–45. https://doi.org/10.1016/j.jtbi.2017.06.011

    Article  Google Scholar 

  28. Lim W, Lee S, Sunwoo M, Jo K (2018) . IEEE Trans Intell Transp Syst 19(2):613–626. https://doi.org/10.1109/TITS.2017.2756099

    Article  Google Scholar 

  29. Dosovitskiy A, Ros G, Codevilla F, Lopez A, Koltun V (2017) In: Proceedings of the 1st annual conference on robot learning, pp 1–16

  30. Niu F, Recht B, Re C, Wright SJ (2011) Hogwild!: A lock-free approach to parallelizing stochastic gradient descent

  31. Lazaridis A, Fachantidis A, Vlahavas I (2020) . J Artif Intell Res 69:1421–1471

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Maria J. P. Peixoto.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Peixoto, M.J.P., Azim, A. Improving environmental awareness for autonomous vehicles. Appl Intell 53, 1842–1854 (2023). https://doi.org/10.1007/s10489-022-03468-6

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-022-03468-6

Keywords

Navigation