Abstract
Inspired by box jellyfish that has distributed and complementary perceptive system, we seek to equip manipulator with a camera and an Inertial Measurement Unit (IMU) to perceive ego motion and surrounding unstructured environment. Before robot perception, a reliable and high-precision calibration between camera, IMU and manipulator is a critical prerequisite. This paper introduces a novel calibration system. First, we seek to correlate the spatial relationship between the sensing units and manipulator in a joint framework. Second, the manipulator moving trajectory is elaborately designed in a spiral pattern that enables full excitations on yaw–pitch–roll rotations and x–y–z translations in a repeatable and consistent manner. The calibration has been evaluated on our collected visual inertial-manipulator dataset. The systematic comparisons and analysis indicate the consistency, precision and effectiveness of our proposed calibration method.
Similar content being viewed by others
References
Sinatra, N. R., Teeple, C. B., Vogt, D. M., Parker, K. K., Gruber, D. F., & Wood, R. J. (2019). Ultragentle manipulation of delicate structures using a soft robotic gripper. Science Robotics, 4, eaax5425.
Ren, Z. Y., Hu, W. Q., Dong, X. G., & Sitti, M. (2019). Multi-functional soft-bodied jellyfish-like swimming. Nature Communication, 10, 1–12.
Kawano, S., Miyagawa, S., & Shirai, A. (2016). Systematical experiment for optimal design of vibrating flow pump with jelly-fish valve. Journal of Bionic Engineering, 13, 166–179.
Gershwin, L. (2017). Spineless: The science of jellyfish and the art of growing a backbone. Nature, 551, 32–32.
Yuan, Y., Ning, H. L., & Lu, X. Q. (2019). Bio-inspired representation learning for visual attention prediction. IEEE Transactions on Cybernetics, 51, 3562–3575.
Li, D. R., Liu, Y., & Yuan, X. X. (2013). Image-based self-position and orientation method for moving platform. Science China-Information Sciences, 56, 042301.
Sun, T. R., Cheng, L., Hou, Z. G., & Tan, M. (2021). Novel sliding-mode disturbance observer-based tracking control with applications to robot manipulators. Science China-Information Sciences, 64, 172205.
Xing, H. M., Shi, L. W., Tang, K., Guo, S. X., Hou, X. H., Liu, Y., Liu, H. K., & Hu, Y. (2019). Robust rgb-d camera and imu fusion-based cooperative and relative close-range localization for multiple turtle inspired amphibious spherical robots. Journal of Bionic Engineering, 16, 442–454.
Xiao, J. H., Xiong, D., Yu, Q. H., Huang, K. H., Lu, H. M., & Zeng, Z. W. (2020). A real-time sliding-window-based visual-inertial odometry for mavs. IEEE Transactions on Industrial Informatics, 16, 4049–4058.
Luo, B., Chen, H. Y., Quan, F. Y., Zhang, S. W., & Liu, Y. H. (2020). Natural feature-based visual servoing for grasping target with an aerial manipulator. Journal of Bionic Engineering, 17, 215–228.
Zou, D. P., Wu, Y. X., Pei, L., Ling, H. B., & Yu, W. X. (2019). Structvio: Visual-inertial odometry with structural regularity of man-made environments. IEEE Transactions on Robotics, 35, 999–1013.
Sa, I., Kamel, M., Burri, M., Bloesch, M., Khanna, R., Popovic, M., Nieto, J., & Siegwart, R. (2018). Build your own visual-inertial drone: A cost-effective and open-source autonomous drone. IEEE Robotics and Automation Magazine, 25, 89–103.
Liu, Y. Z., & Meng, Z. Y. (2020). Online temporal calibration based on modified projection model for visual-inertial odometry. IEEE Transactions on Instrumentation and Measurement, 69, 5197–5207.
Yang, L., Dryanovski, I., Valenti, R. G., Wolberg, G., & Xiao, J. Z. (2020). Rgb-d camera calibration and trajectory estimation for indoor mapping. Autonomous Robots, 44, 1485–1503.
Skogh, C., Garm, A., Nilsson, D. E., & Ekstrm, P. (2006). Bilaterally symmetrical rhopalial nervous system of the box jellyfish tripedalia cystophora. Mrophology, 267, 1391–1405.
Garm, A., Oskarsson, M., & Nilsson, D. E. (2011). Box jellyfish use terrestrial visual cues for navigation. Current Biology, 21, 798–803.
Kim, D., Kang, B. B., Kim, K. B. K., Chol, H., Ha, J., Cho, K. J., & Jo, S. (2019). Eyes are faster than hands: A soft wearable robot learns user intention from the egocentric view. Science Robotics, 4, eaav2949.
Yagi, T., Mangalam, K., Yonetani, R., & Sato, Y. (2018). Future person localization in first-person videos. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, USA, pp. 7593–7602.
Eckenhoff, K., Yang, Y., Geneva, P., & Huang, G. Q. (2019). Tightly-coupled visual-inertial localization and 3-d rigid-body target tracking. IEEE Robotics and Automation Letters, 4, 1541–1548.
Jin, L. Q., Zhang, H., & Ye, C. (2020). Camera intrinsic parameters estimation by visual inertial odometry for a mobile phone with application to assisted navigation. IEEE/ASME Transactions on Mechatronics, 25, 1803–1811.
Santoso, F., Garratt, M. A., & Anavatti, S. G. (2017). Visual-inertial navigation systems for aerial robotics: Sensor fusion and technology. IEEE Transactions on Automation Science and Engineering, 14, 260–275.
Kim, D., Shin, S., & Kweon, I. S. (2018). On-line initialization and extrinsic calibration of an inertial navigation system with a relative preintegration method on manifold. IEEE Transactions on Automation Science and Engineering, 15, 1272–1285.
Kelly, J., & Sukhatme, G. S. (2011). Visual-inertial sensor fusion: Localization, mapping and sensor-to-sensor self-calibration. The International Journal of Robotics Research, 30, 56–79.
Wang, J. & Olson, E. (2016). Apriltag 2: Efficient and robust fiducial detection. IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Korea, pp. 4193–4198.
Lobo, J., & Dias, J. (2007). Relative pose calibration between visual and inertial sensors. The International Journal of Robotics Research, 26, 561–575.
Forster, C., Carlone, L., Dellaert, F., & Scaramuzza, D. (2017). On-manifold preintegration for real-time visual-inertial odometry. IEEE Transactions on Robotics, 33, 1–21.
Tian, F. P., Feng, W., Zhang, Q., Wang, X. W., Sun, J. Z., Loia, V., & Liu, Z. Q. (2019). Active camera relocalization from a single reference image without hand-eye calibration. IEEE Transactions on Pattern Analysis and Machine Intelligence, 41, 2791–2806.
Choi, Y., Kim, N., Hwang, S., Park, K., Yoon, J. S., An, K., & Kweon, I. S. (2018). Kaist multi-spectral day/night data set for autonomous and assisted driving. IEEE Transactions on Intelligent Transportation Systems, 19, 934–948.
Gao, P. Y., Li, K., Song, T. X., & Liu, Z. J. (2018). An accelerometers-size-effect self-calibration method for tri-axis rotational inertial navigation system. IEEE Transactions Industrial Electronics, 65, 1655–1664.
Xu, W. F., Meng, D. S., Liu, H. D., Wang, X. Q., & Liang, B. (2019). Singularity-free trajectory planning of free-floating multi-arm space robots for keeping the base inertially stabilized. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 49, 2464–2477.
Furgale, P., Rehder, J., & Siegwart, R. (2013) Unified temporal and spatial calibration for multi-sensor systems. IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Tokyo, Japan, pp. 1280–1286.
He, H. S., Li, Y., & Tan, J. D. (2016). Rotational coordinate transformation for visual-inertial sensor fusion. International Conference on Social Robotics (ICSR), Qingdao, China, pp. 430–440.
Leutenegger, S., Lynen, S., Bosse, M., Siegwart, R., & Furgale, P. (2015). Keyframe-based visual-inertial odometry using nonlinear optimization. The International Journal of Robotics Research, 34, 314–334.
Li, Y. G., Hou, C., Luo, Y., Zhao, Y. W., & Zhao, X. G. (2021). A arameters identification method for flexible joints based on resonance and anti-resonance frequency characteristics. Robot, 43, 279–288.
Acknowledgements
This work was supported by the National Natural Science Foundation of China (61903357, 61902299, 62022088), the International Partnership Program of Chinese Academy of Sciences (173321KYSB20200002), Liaoning Provincial Natural Science Foundation of China (2020-MS-032, 2021JH6/ 10500114, 2020JH2/10500002), Guangzhou Science and Technology Planning Project (202102021300) and China Postdoctoral Science Foundation (2019TQ0239, 2019M663636).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that there are no conflicts of interest to this work.
Rights and permissions
About this article
Cite this article
Zhang, Y., Liang, W., Zhang, S. et al. High-precision Calibration of Camera and IMU on Manipulator for Bio-inspired Robotic System. J Bionic Eng 19, 299–313 (2022). https://doi.org/10.1007/s42235-022-00163-7
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s42235-022-00163-7