Abstract
With recent technological developments, The UAS (unmanned aerial system) has been recognized for its value and usefulness in various fields. Prior researchers have utilized several drones in collaboration to navigate to achieve common goals such as target tracking, rescue operations, and target-finding with multi-UAS systems. Multi-agent reinforcement learning algorithms are a type of artificial intelligence technology in which many agents collaborate to perform tasks. When a multi-UAS cooperative navigation technique is deployed to a complicated environment such as an urban logistics system, the agents’ learning capacities become more tedious. In this study, we present what is termed the improved Multi-Actor-Attention-Critic (iMAAC) approach, a modified multi-agent reinforcement learning method for application to urban air mobility logistic services. A virtual simulation environment based on Unity is created to validate the suggested method. In the virtual environment, the real-world situation of UAS logistics development services is replicated. When the findings are compared to those of other landmark reinforcement algorithms, iMAAC shows a higher learning rate than those by the other algorithms when utilized in multi-agent systems.
The authors contributed equally to this work
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Delivery drone services market grows at rate of 28% with increased demand for contactless deliveries during COVID-19 [Internet] [cited 2021 Oct 28]. Available from: https://www.einnews.com/pr_news/548404909/delivery-drone-services-market-grows-at-rate-of-28-with-increased-demand-for-contactless-deliveries-during-covid-19
Learn about how Wing delivery works [Internet] [cited 2021 Oct 28]. Available from: https://wing.com/about-delivery/
Iqbal S, Sha F (2019) Actor-attention-critic for multi-agent reinforcement learning. In: 36th International conference on machine learning, ICML 2019
Muñoz G, Barrado C, Çetin E, Salami E (2019) Deep reinforcement learning for drone delivery. Drones 3(3)
Shah S, Dey D, Lovett C, Kapoor A (2017) AirSim: high-fidelity visual and physical simulation for autonomous vehicles. CoRR [Internet]. abs/1705.05065. Available from: http://arxiv.org/abs/1705.05065
Chen X, Ulmer MW, Thomas BW (2021) Deep Q-learning for same-day delivery with vehicles and drones. Eur J Oper Res
Feng L (2020) Reinforcement learning to optimize the logistics distribution routes of unmanned aerial vehicle
Qie H, Shi D, Shen T, Xu X, Li Y, Wang L (2019) Joint optimization of multi-UAV target assignment and path planning based on multi-agent reinforcement learning. IEEE Access 7:146264–146272
Yun WJ, Jung S, Kim J, Kim J-H (2021) Distributed deep reinforcement learning for autonomous aerial eVTOL mobility in drone taxi applications. ICT Express [Internet], 7(1):1–4. Available from: https://www.sciencedirect.com/science/article/pii/S2405959521000059
Unity Real-Time Development Platform|3D, 2D VR & AR Engine [Internet] [cited 2021 Sep 2]. Available from: https://unity.com/
Acknowledgements
This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (No. 2020R1A6A1A03046811).
This work was supported the National Foundation of Korea (NRF) grant funded by the Korea government(MIST) (No. 2021R1A2C209494311).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Jo, H. et al. (2023). Multi-agent Reinforcement Learning-Based UAS Control for Logistics Environments. In: Lee, S., Han, C., Choi, JY., Kim, S., Kim, J.H. (eds) The Proceedings of the 2021 Asia-Pacific International Symposium on Aerospace Technology (APISAT 2021), Volume 2. APISAT 2021. Lecture Notes in Electrical Engineering, vol 913. Springer, Singapore. https://doi.org/10.1007/978-981-19-2635-8_71
Download citation
DOI: https://doi.org/10.1007/978-981-19-2635-8_71
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-2634-1
Online ISBN: 978-981-19-2635-8
eBook Packages: EngineeringEngineering (R0)