Abstract
Terrain analysis is the basis of geological research. However, due to factors such as distance and range, it is often difficult to study the terrain environment in the field. Therefore, researchers can observe and study the terrain by making a three-dimensional terrain model. The 3D terrain model can reduce the terrain range, eliminate the limitation on distance, and control the scene through program interface, to achieve human-computer interaction to meet different research needs. The usual human-computer interaction methods are implemented through traditional peripherals such as the mouse and keyboard. With the rapid development of computer network technology and the continuous improvement of intelligent software and hardware, people have greater requirements for interactive manipulation and immersion. This article proposes a method for displaying terrain models based on real-sensing technology by using Intel’s RealSense camera to control the scene of the desert model through gestures. The user can observe the model from two different perspectives, and use different gestures to zoom in, zoom out, move, and rotate the scene, as well as choose some options. The traditional method of controlling by mouse is also applicable. The entire project is designed as a game, with a realistic and complete model, an exquisite interface, and strong interactivity.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Alisi, T., Bimbo, A.D., Valli, A.: Natural interfaces to enhance visitors’ experiences. IEEE Multimed. 12(3), 80–85 (2005)
Baraldi, S., Bimbo, A.D., Landucci, L., Valli, A.: wikiTable: finger driven interaction for collaborative knowledge-building workspaces. In: Proceedings of 2006 IEEE International Conference on Computer Vision and Pattern Recognition Workshop, p. 144 (2006)
Kaltenbrunner, M., Jordà, S., Geiger, G., Alonso, M.: The reactable: a collaborative musical instrument. In: Workshop on Tangible Interaction in Collaborative Environments (2006)
Kamel, A., Sheng, B., Yang, P., Li, P., Shen, R., Feng, D.D.: Deep convolutional neural networks for human action recognition using depth maps and postures. IEEE Trans. Syst. Man Cybern. Syst. 49(9), 1806–1819 (2019)
Karambakhsh, A., Kamel, A., Sheng, B., Li, P., Yang, P., Feng, D.D.: Deep gesture interaction for augmented anatomy learning. Int. J. Inf. Manage. 45, 328–336 (2019)
Ke, Q., Bennamoun, M., An, S., Sohel, F., Boussaid, F.: A new representation of skeleton sequences for 3D action recognition. In: CVPR, pp. 4570–4579 (2017)
Lu, P., Sheng, B., Luo, S., Jia, X., Wu, W.: Image-based non-photorealistic rendering for realtime virtual sculpting. Multim. Tools Appl. 74(21), 9697–9714 (2015)
Meng, X., et al.: A video information driven football recommendation system. Comput. Electr. Eng. 85, 106699 (2020)
Oreifej, O., Liu, Z.: HON4D: Histogram of oriented 4D normals for activity recognition from depth sequences. In: CVPR, pp. 716–723 (2013)
Siena, F.L., Byrom, B., Watts, P., Breedon, P.: Utilising the intel RealSense camera for measuring health outcomes in clinical research. J. Med. Syst. 42(3), 1–10 (2018). https://doi.org/10.1007/s10916-018-0905-x
Sheng, B., Li, P., Zhang, Y., Mao, L., Chen, C.L.P.: GreenSea: visual soccer analysis using broad learning system. IEEE Trans. Cybern. 1–15 (2020). https://doi.org/10.1109/TCYB.2020.2988792. https://ieeexplore.ieee.org/document/9098099
Stefano, B., Alberto, D., Lea, L., Nicola, T.: Natural interaction. In: Encyclopedia of Database Systems. https://link.springer.com/referenceworkentry/10.1007
Vemulapalli, R., Arrate, F., Chellappa, R.: Human action recognition by representing 3D skeletons as points in a lie group. In: CVPR, pp. 588–595 (2014)
Vemulapalli, R., Chellappa, R.: Rolling rotations for recognizing human actions from 3D skeletal data. In: CVPR, pp. 4471–4479 (2016)
Wang, L., Huynh, D.Q.: Koniusz: a comparative review of recent Kinect-based action recognition algorithms. IEEE Trans. Image Process. 29, 15–28 (2020)
Xia, L., Chen, C., Aggarwal, J. K.: View invariant human action recognition using histograms of 3D joints. In: CVPR, pp. 20–27 (2012)
Yang, X., Tian, Y.: Super normal vector for activity recognition using depth sequences. In: CVPR, pp. 804–811 (2014)
Zhang, P., et al.: Tracking soccer players using spatio-temporal context learning under multiple views. Multimedia Tools Appl. 77(15), 18935–18955 (2018)
Zhu, Y., Fujimura, K.: Constrained optimization for human pose estimation from depth sequences. In: Asian Conference on Computer Vision [S.l.], pp. 408–418. IEEE Press (2007)
Shotton, J., et al.: Real-time human pose recognition in parts from single depth images. In: Computer Vision and Pattern Recognition. [S.l.], pp. 1297–1304. IEEE Press (2011)
Acknowledgement
This work was supported in part by the National Key Research and Development Program of China under Grant 2018YFF0300903, in part by the National Natural Science Foundation of China under Grant 61872241 and Grant 61572316, and in part by the Science and Technology Commission of Shanghai Municipality under Grant 15490503200, Grant 18410750700, Grant 17411952600, and Grant 16DZ0501100.
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Zhang, W., Zhu, F., Lu, P., Li, P., Sheng, B., Mao, L. (2020). 3D Geology Scene Exploring Base on Hand-Track Somatic Interaction. In: Magnenat-Thalmann, N., et al. Advances in Computer Graphics. CGI 2020. Lecture Notes in Computer Science(), vol 12221. Springer, Cham. https://doi.org/10.1007/978-3-030-61864-3_30
Download citation
DOI: https://doi.org/10.1007/978-3-030-61864-3_30
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-61863-6
Online ISBN: 978-3-030-61864-3
eBook Packages: Computer ScienceComputer Science (R0)