Abstract
The ability to recognize and follow human beings is considered to be a key technology for mobile robots. A vision-based following approach is proposed in this paper. It consists of two parts: the human detection part and the visual servoing part. In the human detection part, the image information is captured by the robot camera and processed by a neural network detection algorithm to extract human position and state information. In the visual servoing part, robot switches between three modes including seeking, following and stopping for human-following. Technically, an algorithm is proposed to calculate the angle between the robot and the human, which is used to control. The innovation of this paper is the use of a full vision-based approach to human-following, where the whole process only requires image information. A series of experiments were conducted to verify the effectiveness of the approach.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Jung, E.J., Lee, J.H., Yi, B.J., et al.: Development of a laser-range-finder-based human tracking and control algorithm for a marathoner service robot. IEEE/ASME Trans. Mechatron. 19(6), 1963–1976 (2013)
Zhang, M.Y., Liu, X.L., Xu, D., et al.: Vision-based target-following guider for mobile robot. IEEE Trans. Industr. Electron. 66(12), 9360–9371 (2019)
Feng, T., Yu, Y., Wu, L., et al.: A human-tracking robot using ultra wideband technology. IEEE Access 6, 42541–42550 (2018)
Laneurit, J., Chapuis, R., Debain, C.: TRACKBOD, an accurate, robust and low cost system for mobile robot person following. In: International Conference on Machine Control & Guidance (2016)
Germa, T., Lerasle, F., Ouadah, N., et al.: Vision and RFID data fusion for tracking people in crowds by a mobile robot. Comput. Vis. Image Underst. 114(6), 641–651 (2010)
Zheng, Y.Y., Ge, J.: Binocular intelligent following robot based on YOLO-LITE. In: 2nd International Conference on Computer Science Communication and Network Security, p. 10 (2021)
Tsai, T.H., Yao, C.H.: A robust tracking algorithm for a human-following mobile robot. IET Image Process. 15(3), 786–796 (2021)
Liu, H.L., Luo, J., Wu, P., et al.: People detection and tracking using RGB-D cameras for mobile robots. Int. J. Adv. Robot. Syst. 13(5) (2016). https://doi.org/10.1177/1729881416657746
Han, D., Peng, Y.G.: Human-following of mobile robots based on object tracking and depth vision. In: 3rd International Conference on Mechatronics, Robotics and Automation, pp.105–109 (2020)
Mi, W.M., Wang, X.Z., Ren, P., et al.: A system for an anticipative front human following robot. In: International Conference on Artificial Intelligence and Robotics and the International Conference on Automation, Control and Robotics Engineering, pp. 1–6 (2016)
Sharma, A.K., Pandey, A., Khan, M.A., et al.: Human following robot. In: 2021 International Conference on Advance Computing and Innovative Technologies in Engineering, pp. 440–446 (2021)
Suzuki, S., Mitsukura, Y., Takimoto, H., et al.: A human tracking mobile-robot with face detection. In: 35th Annual Conference of IEEE Industrial Electronics, pp. 4217–4222 (2009)
Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20(3), 273–297 (1995). https://doi.org/10.1007/BF00994018
Girshick, R., Donahue, J., Darrell, T., et al.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580–587 (2014)
Redmon, J., Divvala, S., Girshick, R., et al.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779–788 (2016)
Liu, Z., Lin, Y.T., Cao, Y., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 10012–10022 (2021)
Liu, Z., Mao, H., Wu, C.-Y., et al.: A ConvNet for the 2020s. arXiv preprint arXiv:2201.03545 (2022)
Ge, Z., Liu, S., Wang, F., Li, Z., et al.: YOLOX: exceeding yolo series in 2021. arXiv preprint arXiv:2107.08430 (2021)
Jocher, G., et al.: YOLOv5 (2021). https://github.com/ultralytics/yolov5
Lin, T.-Y., Dollar, P., Girshick, R., et al.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 936–944 (2017)
Liu, S., Qi, L., Qin, H.F., et al.: Path aggregation network for instance segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8759–8768 (2018)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Tan, Y., Wu, S., Liu, W., Yang, X., Sun, HJ. (2022). A Monocular Vision-Based Human-Following Approach for Mobile Robots. In: Liu, H., et al. Intelligent Robotics and Applications. ICIRA 2022. Lecture Notes in Computer Science(), vol 13458. Springer, Cham. https://doi.org/10.1007/978-3-031-13841-6_65
Download citation
DOI: https://doi.org/10.1007/978-3-031-13841-6_65
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-13840-9
Online ISBN: 978-3-031-13841-6
eBook Packages: Computer ScienceComputer Science (R0)