Skip to main content
Log in

Reference cage architecture for autonomous docking of mobile robots in automotive production systems

  • ORIGINAL ARTICLE
  • Published:
The International Journal of Advanced Manufacturing Technology Aims and scope Submit manuscript

Abstract

This paper addresses the need for high-accuracy docking in material handling processes using autonomous mobile robots in factories. To satisfy this need for the tasks, such as loading, unloading, and reaching the charging station, traditional navigation methods often rely on physical restraints at the docking stations, which limits flexibility in production lines. To achieve high-level accuracy without such restrictions, this study proposes the reference cage architecture, which utilizes multi-reference points to maintain scan-matching-based localization performance during docking. The contributions of this research include achieving sub-centimeter accuracy in pose estimation near the target pose and the development of a real-time reference selection decision mechanism. To verify the effectiveness of the proposed approach, extensive testing and validation have been conducted on the automotive production lines of the Ford Otosan Golcuk Plant. These tests consider real-world operational conditions, such as noises, disturbances, and outliers, setting this study apart from similar publications in the literature. The results demonstrate the potential of the reference cage architecture in enabling high-accuracy docking in autonomous mobile robot applications within factory environments.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Algorithm 1
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14

Similar content being viewed by others

Availability of data and materials

The authors declare that the data supporting the findings are available online at https://osf.io/ke4n9/ on The Open Science Framework (OSF).

Code Availability

The binary packages used in this research are available as ROS repositories at https://packages.ros.org/ros/ubuntu/. The rest of the code and scripts are reproducible as discussed in this article.

Notes

  1. https://www.fordotosan.com.tr/en/operations/production/plants

  2. https://github.com/yilmazabdurrah/synthetic_laser_measurement_MATLAB

  3. http://wiki.ros.org/vrpn_client_ros

References

  1. Gharebaghi A, Abolfazl Mostafavi M, Larouche C, Esmaeili K, Genon M (2022) Precise indoor localization and mapping using mobile laser scanners: a scoping review. Geomatica 75(4):165–177. https://doi.org/10.1139/geomat-2021-0011

    Article  Google Scholar 

  2. Alatise MB, Hancke GP (2020) A review on challenges of autonomous mobile robot and sensor fusion methods. IEEE Access 8:39830–39846. https://doi.org/10.1109/ACCESS.2020.2975643

    Article  Google Scholar 

  3. Tao B, Wu H, Gong Z, Yin Z, Ding H (2020) An RFID-based mobile robot localization method combining phase difference and readability. IEEE Trans Autom Sci Eng 18(3):1406–1416. https://doi.org/10.1109/TASE.2020.3006724

    Article  Google Scholar 

  4. Tripicchio P, DAvella S, Unetti M, (2022) Efficient localization in warehouse logistics: a comparison of LMS approaches for 3D multilateration of passive UHF RFID tags. Int J Adv Manuf Technol 120(7–8):4977–4988. https://doi.org/10.1007/s00170-022-09018-1

  5. Li X, Du S, Li G, Li H (2019) Integrate point-cloud segmentation with 3d lidar scan-matching for mobile robot localization and mapping. Sensors 20(1):237. https://doi.org/10.3390/s20010237

    Article  MathSciNet  Google Scholar 

  6. Shu M, Chen G, Zhang Z (2021) 3D point cloud-based indoor mobile robot in 6-dof pose localization using a wi-fi-aided localization system. IEEE Access 9:38636–38648. https://doi.org/10.1109/ACCESS.2021.3060760

    Article  Google Scholar 

  7. Alapetite A, Wang Z, Hansen JP, Zajaczkowski M, Patalan M (2020) Comparison of three off-the-shelf visual odometry systems. Robotics 9(3):56. https://doi.org/10.3390/robotics9030056

    Article  Google Scholar 

  8. Chen X-X, Huang J (2019) Odor source localization algorithms on mobile robots: a review and future outlook. Robot Auton Syst 112:123–136. https://doi.org/10.1016/j.robot.2018.11.014

    Article  Google Scholar 

  9. Li C, Wang S, Zhuang Y, Yan F (2019) Deep sensor fusion between 2D laser scanner and IMU for mobile robot localization. IEEE Sens J 21(6):8501–8509. https://doi.org/10.1109/JSEN.2019.291082

    Article  Google Scholar 

  10. Nemec D, Šimák V, Janota A, Hruboš M, Bubeníková E (2019) Precise localization of the mobile wheeled robot using sensor fusion of odometry, visual artificial landmarks and inertial sensors. Robot Auton Syst 112:168–177. https://doi.org/10.1016/j.robot.2018.11.019

    Article  Google Scholar 

  11. Savci IH, Yilmaz A, Karaman S, Ocakli H, Temeltas H (2022) Improving navigation stack of a ros-enabled industrial autonomous mobile robot (amr) to be incorporated in a large-scale automotive production. Int J Adv Manuf Technol 120(5–6):3647–3668. https://doi.org/10.1007/s00170-022-08883-0

    Article  Google Scholar 

  12. Nguyen TT, Ngo DH, Nguyen QT, Tran DT, Dang XB (2022) Indoor mobile robot positioning using sensor fusion. In: 2022 6th international conference on green technology and sustainable development (GTSD). IEEE, pp 884–890. https://doi.org/10.1109/GTSD54989.2022.9989084

  13. Baharom AK, Abdul-Rahman S, Jamali R, Mutalib S (2020) Towards modelling autonomous mobile robot localization by using sensor fusion algorithms. In: 2020 IEEE 10th international conference on system engineering and technology (ICSET). IEEE, pp 185–190. https://doi.org/10.1109/ICSET51301.2020.9265372

  14. Cheng J, Zhang H (2020) Meng MQ-H: improving visual localization accuracy in dynamic environments based on dynamic region removal. IEEE Trans Autom Sci Eng 17(3):1585–1596. https://doi.org/10.1109/TASE.2020.2964938

    Article  Google Scholar 

  15. Zhang T, Zhang H, Li X, Chen J, Lam TL, Vijayakumar S (2021) Acousticfusion: fusing sound source localization to visual slam in dynamic environments. In: 2021 IEEE/RSJ international conference on intelligent robots and systems (IROS). IEEE, pp 6868–6875 https://doi.org/10.1109/IROS51168.2021.9636585

  16. Ran T, Yuan L, Zhang J, Tang D, He L (2021) RS-SLAM: a robust semantic SLAM in dynamic environments based on RGB-D sensor. IEEE Sens J 21(18):20657–20664. https://doi.org/10.1109/JSEN.2021.3099511

    Article  Google Scholar 

  17. Cui L, Ma C (2019) SOF-SLAM: a semantic visual SLAM for dynamic environments. IEEE Access 7:166528–166539. https://doi.org/10.1109/ACCESS.2019.2952161

    Article  Google Scholar 

  18. Panigrahi PK, Bisoy SK (2022) Localization strategies for autonomous mobile robots: a review. J King Saud Univ - Comput Inf Sci 34(8):6019–6039. https://doi.org/10.1016/j.jksuci.2021.02.015

    Article  Google Scholar 

  19. Yilmaz A, Temeltas H (2019) Self-adaptive Monte Carlo method for indoor localization of smart AGVs using LIDAR data. Robot Auton Syst 122. https://doi.org/10.1016/j.robot.2019.103285

  20. Vasiljević G, Miklić D, Draganjac I, Kovačić Z, Lista P (2016) High-accuracy vehicle localization for autonomous warehousing. Robot Comput -Integr Manuf 42:1–16. https://doi.org/10.1016/j.rcim.2016.05.001

    Article  Google Scholar 

  21. Yilmaz A, Sumer E, Temeltas H (2022) A precise scan matching based localization method for an autonomously guided vehicle in smart factories. Robot Comput -Integr Manuf 75:102302. https://doi.org/10.1016/j.rcim.2021.102302

    Article  Google Scholar 

  22. Liu Y, Wang C, Wu H, Wei Y, Ren M, Zhao C (2022) Improved lidar localization method for mobile robots based on multi-sensing. Remote Sensing 14(23):6133. https://doi.org/10.3390/rs14236133

    Article  Google Scholar 

  23. Maken FA, Ramos F, Ott L (2022) Bayesian iterative closest point for mobile robot localization. Int J Rob Res 41(9–10):851–874. https://doi.org/10.1177/02783649221101417

    Article  Google Scholar 

  24. Lasi H, Fettke P, Kemper H-G, Feld T, Hoffmann M (2014) Industry 4.0. Bus Inf Syst Eng 6(4):239–242. https://doi.org/10.1007/s12599-014-0334-4

    Article  Google Scholar 

  25. Bouraine S, Bougouffa A, Azouaoui O (2022) Particle swarm optimization for solving a scan-matching problem based on the normal distributions transform. Evol Intel 15(1):683–694. https://doi.org/10.1007/s12065-020-00545-y

    Article  Google Scholar 

  26. Palieri M, Morrell B, Thakur A, Ebadi K, Nash J, Chatterjee A, Kanellakis C, Carlone L, Guaragnella C, Agha-Mohammadi A-A (2020) Locus: a multi-sensor lidar-centric solution for high-precision odometry and 3d mapping in real-time. IEEE Robot Autom Lett 6(2):421–428. https://doi.org/10.1109/LRA.2020.3044864

    Article  Google Scholar 

  27. Chen K, Lopez BT, Agha-mohammadi A-A, Mehta A (2022) Direct lidar odometry: fast localization with dense point clouds. IEEE Robot Autom Lett 7(2):2000–2007. https://doi.org/10.1109/LRA.2022.3142739

    Article  Google Scholar 

  28. Filotheou A (2022) Correspondenceless scan-to-map-scan matching of homoriented 2d scans for mobile robot localisation. Robot Auton Syst 149:103957. https://doi.org/10.1016/j.robot.2021.103957

    Article  Google Scholar 

  29. Chang C-Y, Wu C-L, Cheng J-M, Jian S-J (2023) Autonomous mobile robots for recycling metal shaving at CNC factories. Int J Adv Manuf Technol 126(5):2205–2218. https://doi.org/10.1007/s00170-023-11284-6

    Article  Google Scholar 

  30. Dellenbach P, Deschaud J.-E, Jacquet B, Goulette F (2022) CT-ICP: real-time elastic LiDAR odometry with loop closure. In: 2022 international conference on robotics and automation (ICRA). IEEE, pp 5580–5586 https://doi.org/10.1109/ICRA46639.2022.9811849

  31. Li L, Kong X, Zhao X, Li W, Wen F, Zhang H, Liu Y (2021) SA-LOAM: semantic-aided LiDAR SLAM with loop closure. In: 2021 IEEE international conference on robotics and automation (ICRA). IEEE, pp 7627–7634

  32. Djordjević V, Stojanović V, Pršić D, Dubonjić L, Morato MM (2022) Observer-based fault estimation in steer-by-wire vehicle. Engineering Today 1(1):7–17. https://doi.org/10.5937/engtoday2201007D

  33. Lee SH, Oh JH, An YC (2021) A new range-only measurement-based glass line feature extraction method. Electron Lett 57(21):804–806. https://doi.org/10.1049/ell2.12268

    Article  Google Scholar 

  34. Kim J, Chung W (2016) Localization of a mobile robot using a laser range finder in a glass-walled environment. IEEE Trans Industr Electron 63(6):3616–3627. https://doi.org/10.1109/TIE.2016.2523460

    Article  Google Scholar 

  35. Yilmaz A, Temeltas H (2021) Integration of affine ICP into the precise localization problem of smart-AGVs: procedures, enhancements and challenges. Trans Inst Meas Control 43(8):1695–1709. https://doi.org/10.1177/0142331220933430

    Article  Google Scholar 

  36. Chetverikov D, Stepanov D, Krsek P (2005) Robust Euclidean alignment of 3D point sets: the trimmed iterative closest point algorithm. Image Vis Comput 23(3):299–309. https://doi.org/10.1016/j.imavis.2004.05.007

    Article  Google Scholar 

  37. He P, Wen J, Stojanovic V, Liu F, Luan X (2022) Finite-time control of discrete-time semi-Markov jump linear systems: a self-triggered MPC approach. J Franklin Inst 359(13):6939–6957. https://doi.org/10.1016/j.jfranklin.2022.06.043

    Article  MathSciNet  MATH  Google Scholar 

  38. Zhou C, Tao H, Chen Y, Stojanovic V, Paszke W (2022) Robust point-to-point iterative learning control for constrained systems: a minimum energy approach. Int J Robust Nonlinear Control 32(18):10139–10161. https://doi.org/10.1002/rnc.6354

    Article  MathSciNet  Google Scholar 

  39. Lee S, Kim H, Lee B (2019) An efficient rescue system with online multi-agent SLAM framework. Sensors 20(1):235. https://doi.org/10.3390/s20010235

    Article  Google Scholar 

  40. Zhang J, Singh S (2017) Low-drift and real-time lidar odometry and mapping. Auton Robot 41:401–416. https://doi.org/10.1007/s10514-016-9548-2

    Article  Google Scholar 

  41. Ulaş C, Temeltaş H (2013) 3d multi-layered normal distribution transform for fast and long range scan matching. J Intell Robot Syst 71:85–108. https://doi.org/10.1007/s10846-012-9780-8

    Article  Google Scholar 

  42. Ram P, Sinha K (2019) Revisiting kd-tree for nearest neighbor search. In: Proceedings of the 25th Acm Sigkdd international conference on knowledge discovery & data mining. p 1378–1388 https://doi.org/10.1145/3292500.3330875

  43. Sutopo H (2011) Bresenham’s lines algorithm visualization using flash. Int J Comput Theory Eng 3(3):422

    Article  Google Scholar 

  44. Du S, Xu G, Zhang S, Zhang X, Gao Y, Chen B (2020) Robust rigid registration algorithm based on pointwise correspondence and correntropy. Pattern Recogn Lett 132:91–98. https://doi.org/10.1016/j.patrec.2018.06.028

    Article  Google Scholar 

Download references

Acknowledgements

We express our gratitude for the ongoing support provided by the Ford Otosan Light Mobility Laboratory at the Gölcük R &D Center and the Robotics Laboratory of the Control and Automation Engineering Department at Istanbul Technical University. Moreover, we present our appreciation to ChatGPT for its support in enhancing the grammar and proofreading of this manuscript.

Funding

This research was financially supported by the Gölcük R &D Center at Ford Otosan.

Author information

Authors and Affiliations

Authors

Contributions

All authors made significant contributions to the structure and design of the manuscript. Abdurrahman Yilmaz developed and implemented the reference cage method. Aycan Deniz Vit and Hakan Ocakli were responsible for material preparation, field tests, data collection, and analysis. Ismail Hakki Savci and Hakan Temeltas provided guidance throughout the project. The initial draft of the manuscript was written by Abdurrahman Yilmaz and Aycan Deniz Vit, with input and feedback from all authors on earlier versions. The final manuscript was reviewed and approved by all authors.

Corresponding author

Correspondence to Abdurrahman Yilmaz.

Ethics declarations

Ethics approval

Not applicable

Consent to participate

Not applicable

Consent for publication

Not applicable

Conflict of interest

The authors declare no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix A: Ground truth reference system for field tests

Appendix A: Ground truth reference system for field tests

Field tests were performed in a 5 m \(\times \) 5 m \(\times \) 3 m volume, covered by OptiTrack motion capture (MoCap) system with four OptiTrack Prime 17W cameras providing the 6D pose of the AMR at 120 Hz. Thanks to this optical camera system, the robot’s position can be tracked with sub-millimeter precision at high frequency. The estimated pose data by the MoCap system is transferred to the ROS environment employing the ROS Virtual Reality Peripheral Network (VRPN) client node.Footnote 3 The overall system architecture is demonstrated in Appendix Fig. 15.

Fig. 15
figure 15

Ground truth reference data generation system overview—Optitrack MoCap system and ROS integration

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Yilmaz, A., Vit, A.D., Savci, I.H. et al. Reference cage architecture for autonomous docking of mobile robots in automotive production systems. Int J Adv Manuf Technol 129, 3497–3511 (2023). https://doi.org/10.1007/s00170-023-12456-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00170-023-12456-0

Keywords

Navigation