Abstract
The existing virtual experiment platform mainly uses virtual reality technology or animation technology to assist students in experimental teaching, but it lacks the standardized supervision of users’ experimental behaviors. To address the above problems, this paper designs a prototype smart glove application for middle school experimental scenarios and proposes a scene perception algorithm based on the smart glove, so as to obtain the user’s experimental behavior more accurately. Based on the perception of the experimental scene, this paper also proposes a multimodal fusion of intelligent navigation interaction paradigm to obtain the user’s experimental intention, thus allowing students to conduct exploratory experiments on a virtual experimental platform with targeted guidance and monitoring of user behavior. Experiments show that the smart glove designed in this paper can sense the relative relationship between experimental equipment and objects in the scene in real time. Based on the user’s experimental behavior, the smart glove can also infer the operator’s experimental intent and provide timely feedback and guidance on the user’s experimental behavior.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Mur-Artal, R., Tardós, J.D.: ORB-SLAM2: an open-source SLAM system for monocular, stereo, and RGB-D cameras. IEEE Trans. Robot. (2017)
Redmon, J., Farhadi, A.: YOLOv3: An Incremental Improvement. arXiv e-prints (2018)
Han, T., Kennington, C., Schlangen, D.: Placing objects in gesture space: toward real-time understanding of spatial descriptions. In: AAAI conference of Artificial Intelligence (2018)
Yongda, D., Fang, L., Huang, X.: Research on multimodal human-robot interaction based on speech and gesture. Comput. Electr. Eng. 72, 443–454 (2018)
Ismail, A.W., Billinghurst, M., Sunar, M.S., Yusof, C.S.: Designing an augmented reality multimodal interface for 6DOF manipulation techniques. In: Arai, K., Kapoor, S., Bhatia, R. (eds.) IntelliSys. AISC, vol. 868, pp. 309–322. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-01054-6_22
Jung, J., et al.: Voice+Tactile: augmenting in-vehicle voice user interface with tactile touchpad interaction voice user interface; tactile feedback touchpad; in-vehicle user interface. In: 2020 CHI Conference on Human Factors in Computing Systems (CHI 2020) (2020)
Sidenmark, L., Gellersen, H.: Eye, head and torso coordination during gaze shifts in virtual reality. ACM Trans. Comput. Hum. Interact. 27(1), 1–40 (2019)
Sullivan, S., et al.: Middle school students’ learning of mechanics concepts through engagement in different sequences of physical and virtual experiments. Int. J. Sci. Educ.
Gabunilas, L., et al.: Utilizing Portable Virtual Reality in Teaching Chemistry (2018)
Fidan, M., Tuncel, M.: Integrating augmented reality into problem based learning: the effects on learning achievement and attitude in physics education. Comput. Educ. 142, 103635.1-1036351.9 (2019)
Yip, J., et al.: Improving quality of teaching and learning in classes by using augmented reality video. Comput. Educ. 128, 88–101 (2018)
Salinas, P., Pulido, R.: Visualization of conics through augmented reality. Procedia Comput. Sci. 75, 147–150 (2015)
Carlos, C.A., et al.: An approach of training virtual environment for teaching electro-pneumatic systems - ScienceDirect. IFAC-PapersOnLine 52(9), 278–284 (2019)
Liu, D., et al.: Integration of virtual labs into science E-learning. Procedia Comput. Sci. 75(2015), 95–102 (2015)
Hodges, G.W., et al.: An exploratory study of blending the virtual world and the laboratory experience in secondary chemistry classrooms. Comput. Educ. 122, 179–193 (2018)
Aljuhani, K., Sonbul, M., Althabiti, M., Meccawy, M.: Creating a Virtual Science Lab (VSL): the adoption of virtual labs in Saudi schools. Smart Learn. Environ. 5(1), 1–13 (2018). https://doi.org/10.1186/s40561-018-0067-9
Sotomayor-Moriano, J., Pérez, G., Soto, M.: A Virtual laboratory environment for control design of a multivariable process. IFAC-PapersOnLine 52(9), 15–20 (2019)
Acknowledgment
This paper is supported by the National Key R&D Program of China (No. 2018YFB1004901), and the Independent Innovation’ Team Project of Jinan City (No. 2019GXRC013).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Wang, H., Meng, X., Feng, Z. (2022). Research on the Structure and Key Algorithms of Smart Gloves Oriented to Middle School Experimental Scene Perception. In: Sun, Y., et al. Computer Supported Cooperative Work and Social Computing. ChineseCSCW 2021. Communications in Computer and Information Science, vol 1491. Springer, Singapore. https://doi.org/10.1007/978-981-19-4546-5_32
Download citation
DOI: https://doi.org/10.1007/978-981-19-4546-5_32
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-4545-8
Online ISBN: 978-981-19-4546-5
eBook Packages: Computer ScienceComputer Science (R0)