Abstract
Visualizing the local environment for remote people is necessary for augmented reality (AR) and remote people between local and remote participants. Three dimensional (3D) reconstruction is a promising method for capturing the shapes and appearances in a local space, but accurate reconstruction requires high-precision sensors and high-performance computers. However, general-purpose smartphones and tablets have distance sensors like LiDAR, and we realized a meta-AR space based on such devices. Through synchronized AR and virtual spaces, local and remote users can collaborate in our meta-AR space. The performance and coverage of a smartphone-based 3D reconstruction system are two issues, though. First, 3D reconstruction using a conventional ray cast is slow; however, using a depth map, improvement is possible. Second, users are unable to see the captured space’s coverage; however, doing so is useful. The system cannot adequately reconstruct the space when the coverage is too low. This paper describes a quick 3D reconstruction system that supports the reconstruction with a depth map and mapping coverage visualization. Our experimental findings show its effectiveness, mapping coverage, and utility. The mapping coverage is adequate and our proposed method is 258 times faster than the ray cast method. Moreover, our mapping-coverage visualization increased the texture by 15%.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Corti, A., Giancola, S., Mainetti, G., Sala, R.: A metrological characterization of the kinect v2 time-of-flight camera. Robot. Auton. Syst. 75, 584–594 (2016)
Izadi, S., et al.: Kinect fusion: real-time 3D reconstruction and interaction using a moving depth camera. In: Proceedings of the 24th Annual ACM Symposium on User Interface Software and Technology, pp. 559–568 (2011)
Pagliari, D., Menna, F., Roncella, R., Remondino, F., Pinto, L.: Kinect fusion improvement using depth camera calibration. Int. Arch. Photogrammetry Remote Sens. Spat. Inf. Sci. 45, 479–485 (2014)
Zingoni, A., Diani, M., Corsini, G., Masini, A.: Real-time 3D reconstruction from images taken from an UAV. Int. Arch. Photogrammetry Remote Sens. Spat. Inf. Sci. 40, 313–319 (2015)
Wu, S., Rupprecht, C., Vedaldi, A.: Unsupervised learning of probably symmetric deformable 3D objects from images in the wild. In: Proceedings of the CVPR, pp. 1–10 (2020)
Stotko, P., Krunnpen, S., Hullin, M., Weinnnann, M., Klein, R.: SLAMCast: large-scale, real-time 3D reconstruction and streaming for immersive multi-client live telepresence. IEEE Trans. Visual Comput. Graphics 25(5), 2102–2112 (2019)
Zhu, Z., Liu, C., Xu, X.: Visualisation of the digital twin data in manufacturing by using augmented reality. CMS 81, 898–903 (2019)
Obermair, F., et al.: Maintenance with augmented reality remote support in comparison to paper-based instructions: experiment and analysis. In: Proceedings of the ICIEA, pp. 942–947 (2020)
Gerstweiler, G., Vonach, E., Kaufmann, H.: HyMoTrack: a mobile AR navigation system for complex indoor environments. Sensors 16(1), 1–19 (2016)
Rajeev, S., Samani, A., Panetta, K., Agaian, S.: 3D navigational insight using AR technology. In: Proceedings of the HST, pp. 1–4 (2019)
Aschauer, A., Reisner-Kollmann, I., Wolfartsberger, J.: Creating an open-source augmented reality remote support tool for industry: challenges and learnings. Proc. ISM 180, 269–279 (2021)
Anton, D., Kurillo, G., Bajcsy, R.: User experience and interaction performance in 2D/3D telecollaboration. Futur. Gener. Comput. Syst. 82, 77–88 (2018)
Cao, D., Kim, S.: Immediate presence of handwritten annotation on moving picture in real-time collaborative system. IEEE Access 8, 133258–133268 (2020)
Wang, Y., Wang, P., Luo, Z., Yan, Y.: A novel AR remote collaborative platform for sharing 2.5D gestures and gaze. Int. J. Adv. Manufact. Technol. 119(910), 6413–6421 (2022)
Yamada, K., Kimura, A.: A performance evaluation of keypoints detection methods SIFT and AKAZE for 3D reconstruction. In: Proceedings of the IWAIT, pp. 1–4 (2018)
Acknowledgements
This work was supported in part by JSPS KAKENHI Grant Numbers JP19K12266, JP22K18006.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Yasue, K., Kikuchi, M., Ozono, T. (2023). Developing a Meta-AR Space Construction System and Mapping Coverage Visualization. In: Selvaraj, H., Fujimoto, T. (eds) Applied Systemic Studies. ICSEng 2022. Lecture Notes in Networks and Systems, vol 611. Springer, Cham. https://doi.org/10.1007/978-3-031-27470-1_21
Download citation
DOI: https://doi.org/10.1007/978-3-031-27470-1_21
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-27469-5
Online ISBN: 978-3-031-27470-1
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)