Skip to main content
Log in

Evaluation of calibration methods to construct a 3-D environmental map with good color projection using both camera images and laser scanning data

  • Original Article
  • Published:
Artificial Life and Robotics Aims and scope Submit manuscript

Abstract

In mobile robot navigation, restoration of the environment around the robot in a 3-D map is necessary for self-location, route planning, and detecting surrounding obstacles. We construct the 3-D color environmental map using both camera images and laser scanner (LIDAR) point cloud data. In the map, the RGB values are provided to the LIDAR point cloud data by projecting the point cloud onto the simultaneously acquired image. The projection parameters can be determined by measuring the calibration boards using both the camera and the LIDAR. In this paper, we have constructed a method to evaluate the accuracy of projection applicable to any calibration methods. And, then, we found that the calibration points in the central position of an image are important to obtain good projection parameters and that additional points at side positions also can improve the accuracy of the projection.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

References

  1. Yamada R, Yaguchi Y, Yoshida M, Kobayashi S (2018) Towards a system for analyzing accidents of unmanned aerial vehicles. Artif Life Robot 24(1):94–99

    Article  Google Scholar 

  2. Engel J, Schöps T, Cremers D (2014) LSD-SLAM: large-scale direct monocular SLAM. In: European conference on computer vision (ECCV), 6–12 September, pp 1–16

  3. Zhang J, Singh S (2014), LOAM: lidar odometry and mapping in real time. In: Robotics: science and system conference (RSS), 12–16 July, pp 1–19

  4. Veitch-Michaels J, Muller J-P, Walton D, Storey J, Foster M, Crutchely B (2016) Enhancement of stereo imagery by artificial texture projection generated using a LIDAR. Int Arch Photo-gramm Remote Sens Spatial Inf Sci XLI-B5:599–606

    Article  Google Scholar 

  5. De Silva V, Roche J, Kondoz A (2018) Robust fusion of LiDAR and wide-angle camera data for autonomous mobile robots. Sens (Basel) 18(8):E2730. https://doi.org/10.3390/s18082730

    Article  Google Scholar 

  6. Kassir A, Peynot T (2010) Reliable automatic camera-laser calibration, In: Proceeding of the 2010 Australasian conference on robotics and automation, ACRA, December 2010

  7. Park Y, Yun S, Won Chee S, Cho K, Um K, Sim S (2014) Calibration between color camera and 3D LIDAR instruments with a polygonal planar board. Sensors 14:5333–5353

    Article  Google Scholar 

  8. Velas M, Spanel M, Materna, Z, Herout A (2014) Calibration of RGB camera with Velodyne LiDAR. In: WSCG 2014 communication papers proceedings, pp 1–10

  9. Fischler MA, Bolles RC (1981) Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun ACM 24(6):381–395

    Article  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ryuhei Yamada.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Yamada, R., Yaguchi, Y. Evaluation of calibration methods to construct a 3-D environmental map with good color projection using both camera images and laser scanning data. Artif Life Robotics 25, 434–439 (2020). https://doi.org/10.1007/s10015-020-00594-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10015-020-00594-7

Keywords

Navigation