Abstract
This paper addresses the need for high-accuracy docking in material handling processes using autonomous mobile robots in factories. To satisfy this need for the tasks, such as loading, unloading, and reaching the charging station, traditional navigation methods often rely on physical restraints at the docking stations, which limits flexibility in production lines. To achieve high-level accuracy without such restrictions, this study proposes the reference cage architecture, which utilizes multi-reference points to maintain scan-matching-based localization performance during docking. The contributions of this research include achieving sub-centimeter accuracy in pose estimation near the target pose and the development of a real-time reference selection decision mechanism. To verify the effectiveness of the proposed approach, extensive testing and validation have been conducted on the automotive production lines of the Ford Otosan Golcuk Plant. These tests consider real-world operational conditions, such as noises, disturbances, and outliers, setting this study apart from similar publications in the literature. The results demonstrate the potential of the reference cage architecture in enabling high-accuracy docking in autonomous mobile robot applications within factory environments.
Similar content being viewed by others
Availability of data and materials
The authors declare that the data supporting the findings are available online at https://osf.io/ke4n9/ on The Open Science Framework (OSF).
Code Availability
The binary packages used in this research are available as ROS repositories at https://packages.ros.org/ros/ubuntu/. The rest of the code and scripts are reproducible as discussed in this article.
References
Gharebaghi A, Abolfazl Mostafavi M, Larouche C, Esmaeili K, Genon M (2022) Precise indoor localization and map** using mobile laser scanners: a sco** review. Geomatica 75(4):165–177. https://doi.org/10.1139/geomat-2021-0011
Alatise MB, Hancke GP (2020) A review on challenges of autonomous mobile robot and sensor fusion methods. IEEE Access 8:39830–39846. https://doi.org/10.1109/ACCESS.2020.2975643
Tao B, Wu H, Gong Z, Yin Z, Ding H (2020) An RFID-based mobile robot localization method combining phase difference and readability. IEEE Trans Autom Sci Eng 18(3):1406–1416. https://doi.org/10.1109/TASE.2020.3006724
Tripicchio P, DAvella S, Unetti M, (2022) Efficient localization in warehouse logistics: a comparison of LMS approaches for 3D multilateration of passive UHF RFID tags. Int J Adv Manuf Technol 120(7–8):4977–4988. https://doi.org/10.1007/s00170-022-09018-1
Li X, Du S, Li G, Li H (2019) Integrate point-cloud segmentation with 3d lidar scan-matching for mobile robot localization and map**. Sensors 20(1):237. https://doi.org/10.3390/s20010237
Shu M, Chen G, Zhang Z (2021) 3D point cloud-based indoor mobile robot in 6-dof pose localization using a wi-fi-aided localization system. IEEE Access 9:38636–38648. https://doi.org/10.1109/ACCESS.2021.3060760
Alapetite A, Wang Z, Hansen JP, Zajaczkowski M, Patalan M (2020) Comparison of three off-the-shelf visual odometry systems. Robotics 9(3):56. https://doi.org/10.3390/robotics9030056
Chen X-X, Huang J (2019) Odor source localization algorithms on mobile robots: a review and future outlook. Robot Auton Syst 112:123–136. https://doi.org/10.1016/j.robot.2018.11.014
Li C, Wang S, Zhuang Y, Yan F (2019) Deep sensor fusion between 2D laser scanner and IMU for mobile robot localization. IEEE Sens J 21(6):8501–8509. https://doi.org/10.1109/JSEN.2019.291082
Nemec D, Šimák V, Janota A, Hruboš M, Bubeníková E (2019) Precise localization of the mobile wheeled robot using sensor fusion of odometry, visual artificial landmarks and inertial sensors. Robot Auton Syst 112:168–177. https://doi.org/10.1016/j.robot.2018.11.019
Savci IH, Yilmaz A, Karaman S, Ocakli H, Temeltas H (2022) Improving navigation stack of a ros-enabled industrial autonomous mobile robot (amr) to be incorporated in a large-scale automotive production. Int J Adv Manuf Technol 120(5–6):3647–3668. https://doi.org/10.1007/s00170-022-08883-0
Nguyen TT, Ngo DH, Nguyen QT, Tran DT, Dang XB (2022) Indoor mobile robot positioning using sensor fusion. In: 2022 6th international conference on green technology and sustainable development (GTSD). IEEE, pp 884–890. https://doi.org/10.1109/GTSD54989.2022.9989084
Baharom AK, Abdul-Rahman S, Jamali R, Mutalib S (2020) Towards modelling autonomous mobile robot localization by using sensor fusion algorithms. In: 2020 IEEE 10th international conference on system engineering and technology (ICSET). IEEE, pp 185–190. https://doi.org/10.1109/ICSET51301.2020.9265372
Cheng J, Zhang H (2020) Meng MQ-H: improving visual localization accuracy in dynamic environments based on dynamic region removal. IEEE Trans Autom Sci Eng 17(3):1585–1596. https://doi.org/10.1109/TASE.2020.2964938
Zhang T, Zhang H, Li X, Chen J, Lam TL, Vijayakumar S (2021) Acousticfusion: fusing sound source localization to visual slam in dynamic environments. In: 2021 IEEE/RSJ international conference on intelligent robots and systems (IROS). IEEE, pp 6868–6875 https://doi.org/10.1109/IROS51168.2021.9636585
Ran T, Yuan L, Zhang J, Tang D, He L (2021) RS-SLAM: a robust semantic SLAM in dynamic environments based on RGB-D sensor. IEEE Sens J 21(18):20657–20664. https://doi.org/10.1109/JSEN.2021.3099511
Cui L, Ma C (2019) SOF-SLAM: a semantic visual SLAM for dynamic environments. IEEE Access 7:166528–166539. https://doi.org/10.1109/ACCESS.2019.2952161
Panigrahi PK, Bisoy SK (2022) Localization strategies for autonomous mobile robots: a review. J King Saud Univ - Comput Inf Sci 34(8):6019–6039. https://doi.org/10.1016/j.jksuci.2021.02.015
Yilmaz A, Temeltas H (2019) Self-adaptive Monte Carlo method for indoor localization of smart AGVs using LIDAR data. Robot Auton Syst 122. https://doi.org/10.1016/j.robot.2019.103285
Vasiljević G, Miklić D, Draganjac I, Kovačić Z, Lista P (2016) High-accuracy vehicle localization for autonomous warehousing. Robot Comput -Integr Manuf 42:1–16. https://doi.org/10.1016/j.rcim.2016.05.001
Yilmaz A, Sumer E, Temeltas H (2022) A precise scan matching based localization method for an autonomously guided vehicle in smart factories. Robot Comput -Integr Manuf 75:102302. https://doi.org/10.1016/j.rcim.2021.102302
Liu Y, Wang C, Wu H, Wei Y, Ren M, Zhao C (2022) Improved lidar localization method for mobile robots based on multi-sensing. Remote Sensing 14(23):6133. https://doi.org/10.3390/rs14236133
Maken FA, Ramos F, Ott L (2022) Bayesian iterative closest point for mobile robot localization. Int J Rob Res 41(9–10):851–874. https://doi.org/10.1177/02783649221101417
Lasi H, Fettke P, Kemper H-G, Feld T, Hoffmann M (2014) Industry 4.0. Bus Inf Syst Eng 6(4):239–242. https://doi.org/10.1007/s12599-014-0334-4
Bouraine S, Bougouffa A, Azouaoui O (2022) Particle swarm optimization for solving a scan-matching problem based on the normal distributions transform. Evol Intel 15(1):683–694. https://doi.org/10.1007/s12065-020-00545-y
Palieri M, Morrell B, Thakur A, Ebadi K, Nash J, Chatterjee A, Kanellakis C, Carlone L, Guaragnella C, Agha-Mohammadi A-A (2020) Locus: a multi-sensor lidar-centric solution for high-precision odometry and 3d map** in real-time. IEEE Robot Autom Lett 6(2):421–428. https://doi.org/10.1109/LRA.2020.3044864
Chen K, Lopez BT, Agha-mohammadi A-A, Mehta A (2022) Direct lidar odometry: fast localization with dense point clouds. IEEE Robot Autom Lett 7(2):2000–2007. https://doi.org/10.1109/LRA.2022.3142739
Filotheou A (2022) Correspondenceless scan-to-map-scan matching of homoriented 2d scans for mobile robot localisation. Robot Auton Syst 149:103957. https://doi.org/10.1016/j.robot.2021.103957
Chang C-Y, Wu C-L, Cheng J-M, Jian S-J (2023) Autonomous mobile robots for recycling metal shaving at CNC factories. Int J Adv Manuf Technol 126(5):2205–2218. https://doi.org/10.1007/s00170-023-11284-6
Dellenbach P, Deschaud J.-E, Jacquet B, Goulette F (2022) CT-ICP: real-time elastic LiDAR odometry with loop closure. In: 2022 international conference on robotics and automation (ICRA). IEEE, pp 5580–5586 https://doi.org/10.1109/ICRA46639.2022.9811849
Li L, Kong X, Zhao X, Li W, Wen F, Zhang H, Liu Y (2021) SA-LOAM: semantic-aided LiDAR SLAM with loop closure. In: 2021 IEEE international conference on robotics and automation (ICRA). IEEE, pp 7627–7634
Djordjević V, Stojanović V, Pršić D, Dubonjić L, Morato MM (2022) Observer-based fault estimation in steer-by-wire vehicle. Engineering Today 1(1):7–17. https://doi.org/10.5937/engtoday2201007D
Lee SH, Oh JH, An YC (2021) A new range-only measurement-based glass line feature extraction method. Electron Lett 57(21):804–806. https://doi.org/10.1049/ell2.12268
Kim J, Chung W (2016) Localization of a mobile robot using a laser range finder in a glass-walled environment. IEEE Trans Industr Electron 63(6):3616–3627. https://doi.org/10.1109/TIE.2016.2523460
Yilmaz A, Temeltas H (2021) Integration of affine ICP into the precise localization problem of smart-AGVs: procedures, enhancements and challenges. Trans Inst Meas Control 43(8):1695–1709. https://doi.org/10.1177/0142331220933430
Chetverikov D, Stepanov D, Krsek P (2005) Robust Euclidean alignment of 3D point sets: the trimmed iterative closest point algorithm. Image Vis Comput 23(3):299–309. https://doi.org/10.1016/j.imavis.2004.05.007
He P, Wen J, Stojanovic V, Liu F, Luan X (2022) Finite-time control of discrete-time semi-Markov jump linear systems: a self-triggered MPC approach. J Franklin Inst 359(13):6939–6957. https://doi.org/10.1016/j.jfranklin.2022.06.043
Zhou C, Tao H, Chen Y, Stojanovic V, Paszke W (2022) Robust point-to-point iterative learning control for constrained systems: a minimum energy approach. Int J Robust Nonlinear Control 32(18):10139–10161. https://doi.org/10.1002/rnc.6354
Lee S, Kim H, Lee B (2019) An efficient rescue system with online multi-agent SLAM framework. Sensors 20(1):235. https://doi.org/10.3390/s20010235
Zhang J, Singh S (2017) Low-drift and real-time lidar odometry and map**. Auton Robot 41:401–416. https://doi.org/10.1007/s10514-016-9548-2
Ulaş C, Temeltaş H (2013) 3d multi-layered normal distribution transform for fast and long range scan matching. J Intell Robot Syst 71:85–108. https://doi.org/10.1007/s10846-012-9780-8
Ram P, Sinha K (2019) Revisiting kd-tree for nearest neighbor search. In: Proceedings of the 25th Acm Sigkdd international conference on knowledge discovery & data mining. p 1378–1388 https://doi.org/10.1145/3292500.3330875
Sutopo H (2011) Bresenham’s lines algorithm visualization using flash. Int J Comput Theory Eng 3(3):422
Du S, Xu G, Zhang S, Zhang X, Gao Y, Chen B (2020) Robust rigid registration algorithm based on pointwise correspondence and correntropy. Pattern Recogn Lett 132:91–98. https://doi.org/10.1016/j.patrec.2018.06.028
Acknowledgements
We express our gratitude for the ongoing support provided by the Ford Otosan Light Mobility Laboratory at the Gölcük R &D Center and the Robotics Laboratory of the Control and Automation Engineering Department at Istanbul Technical University. Moreover, we present our appreciation to ChatGPT for its support in enhancing the grammar and proofreading of this manuscript.
Funding
This research was financially supported by the Gölcük R &D Center at Ford Otosan.
Author information
Authors and Affiliations
Contributions
All authors made significant contributions to the structure and design of the manuscript. Abdurrahman Yilmaz developed and implemented the reference cage method. Aycan Deniz Vit and Hakan Ocakli were responsible for material preparation, field tests, data collection, and analysis. Ismail Hakki Savci and Hakan Temeltas provided guidance throughout the project. The initial draft of the manuscript was written by Abdurrahman Yilmaz and Aycan Deniz Vit, with input and feedback from all authors on earlier versions. The final manuscript was reviewed and approved by all authors.
Corresponding author
Ethics declarations
Ethics approval
Not applicable
Consent to participate
Not applicable
Consent for publication
Not applicable
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix A: Ground truth reference system for field tests
Appendix A: Ground truth reference system for field tests
Field tests were performed in a 5 m \(\times \) 5 m \(\times \) 3 m volume, covered by OptiTrack motion capture (MoCap) system with four OptiTrack Prime 17W cameras providing the 6D pose of the AMR at 120 Hz. Thanks to this optical camera system, the robot’s position can be tracked with sub-millimeter precision at high frequency. The estimated pose data by the MoCap system is transferred to the ROS environment employing the ROS Virtual Reality Peripheral Network (VRPN) client node.Footnote 3 The overall system architecture is demonstrated in Appendix Fig. 15.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Yilmaz, A., Vit, A.D., Savci, I.H. et al. Reference cage architecture for autonomous docking of mobile robots in automotive production systems. Int J Adv Manuf Technol 129, 3497–3511 (2023). https://doi.org/10.1007/s00170-023-12456-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00170-023-12456-0