Abstract
Input devices used nowadays are typically expensive, lack portability, and require a substantial amount of space. However, these limitations can be overcome by using virtual input devices. In this study, we propose a virtual joystick system, which is a type of virtual input device. Our system detects a handheld stick and computes the direction in which the user’s hand moves relative to a user-defined center. The process of the proposed system can be divided into three stages: preprocessing, handheld stick detection, and direction calculation. In each stage, simple operations are performed and a color camera is used. Therefore, users do not need to purchase a high-performance computer or auxiliary devices to use the proposed system. We compare our system’s performance, gauging accuracy, and speed with those produced using a real joystick. The proposed system’s accuracy is competitive and has real-time speed in the laptop environment.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Lee, S., Jung, J., Hong, J., Ryu, J.B., Yang, H.S.: AR paint: a fusion system of a paint tool and AR. In: International Conference on Entertainment Computing, pp. 122–129. Springer, Berlin, Heidelberg (2012)
Rehg, J.M., Kanade, T.: Digiteyes: Vision-based hand tracking for human-computer interaction. In: Proceedings of 1994 IEEE Workshop on Motion of Non-rigid and Articulated Objects, pp. 16–22. IEEE (1994)
Von Hardenberg, C., Bérard, F.: Bare-hand human-computer interaction. In: Proceedings of the 2001 Workshop on Perceptive User Interfaces, pp. 1–8. ACM (2001)
Zhang, Z., Wu, Y., Shan, Y., Shafer, S.: Visual panel: virtual mouse, keyboard and 3D controller with an ordinary piece of paper. In: Proceedings of the 2001 Workshop on Perceptive User Interfaces, pp. 1–8. ACM (2001)
Iannizzotto, G., Villari, M., Vita, L.: Hand tracking for human-computer interaction with graylevel visual glove: turning back to the simple way. In: Proceedings of the 2001 Workshop on Perceptive User Interfaces, pp. 1–7. ACM (2001)
Niikura, T., Hirobe, Y., Cassinelli, A., Watanabe, Y., Komuro, T., Ishikawa, M.: In-air typing interface for mobile devices with vibration feedback. In ACM SIGGRAPH 2010 Emerging Technologies, p. 15. ACM (2010)
Lee, D., Lee, S.: Vision-based finger action recognition by angle detection and contour analysis. ETRI J. 33(3), 415–422 (2011)
Erdem, A., Erdem, E., Yardimci, Y., Atalay, V., Cetin, A.E.: Computer vision based mouse. In: 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 4, pp. IV-4178. IEEE (2002)
Niyazi, K., Kumar, V., Mahe, S., Vyawahare, S.: Mouse simulation using two coloured tapes. Department of Computer Science, University of Pune, India. Int. J. Inf. Sci. Tech. (IJIST) 2 (2012)
Banerjee, A., Ghosh, A., Bharadwaj, K., Saikia, H.: Mouse control using a web camera based on colour detection (2014). arXiv preprint arXiv:1403.4722
Graetzel, C., Fong, T., Grange, S., Baur, C.: A non-contact mouse for surgeon-computer interaction. Technol. Health Care 12(3), 245–257 (2004)
Robertson, P., Laddaga, R., Van Kleek, M.: Virtual mouse vision based interface. In: Proceedings of the 9th International Conference on Intelligent User Interfaces, pp. 177–183. ACM (2004)
Buchmann, V., Violich, S., Billinghurst, M., Cockburn, A.: FingARtips: gesture based direct manipulation in augmented reality. In: Proceedings of the 2nd International Conference on Computer Graphics and Interactive Techniques in Australasia and South East Asia, pp. 212–221. ACM (2004)
Higuchi, M., Komuro, T.: AR typing interface for mobile devices. In: Proceedings of the 12th International Conference on Mobile and Ubiquitous Multimedia, p. 14. ACM (2013)
Wilson, A.D., Cutrell, E.: Flowmouse: a computer vision-based pointing and gesture input device. In: IFIP Conference on Human-Computer Interaction, pp. 565–578. Springer, Berlin, Heidelberg (2005)
Argyros, A.A., Lourakis, M.I.: Vision-based interpretation of hand gestures for remote control of a computer mouse. In: European Conference on Computer Vision, pp. 40–51. Springer, Berlin, Heidelberg (2006)
Wilson, A.D.: Robust computer vision-based detection of pinching for one and two-handed gesture input. In: Proceedings of the 19th Annual ACM Symposium on User Interface Software and Technology, pp. 255–258. ACM (2006)
Habib, H.A., Mufti, M.: Real time mono vision gesture based virtual keyboard system. IEEE Trans. Consum. Electron. 52(4), 1261–1266 (2006)
Du, H., Charbon, E.: A virtual keyboard system based on multi-level feature matching. In: 2008 Conference on Human System Interactions, pp. 176–181. IEEE (2008)
Murase, T., Moteki, A., Ozawa, N., Hara, N., Nakai, T., Fujimoto, K.: Gesture keyboard requiring only one camera. In: Proceedings of the 24th Annual ACM Symposium Adjunct on User Interface Software and Technology, pp. 9–10. ACM (2011)
Yousaf, M.H., Habib, H.A.: Virtual keyboard: real-time finger joints tracking for keystroke detection and recognition. Arab. J. Sci. Eng. 39(2), 923–934 (2014)
Varona, J., Manresa-Yee, C., Perales, F.J.: Hands-free vision-based interface for computer accessibility. J. Netw. Comput. Appl. 31(4), 357–374 (2008)
Wang, X., Qin, K.: A six-degree-of-freedom virtual mouse based on hand gestures. In: 2010 International Conference on Electrical and Control Engineering, pp. 257–260. IEEE (2010)
Pugeault, N., Bowden, R.: Spelling it out: real-time ASL fingerspelling recognition. In: 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), pp. 1114–1119. IEEE (2011)
Ding, D., Cooper, R.A., Spaeth, D.: Optimized joystick controller. In: The 26th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, vol. 2, pp. 4881–4883. IEEE (2004)
http://www.softpedia.com/get/System/System-Miscellaneous/Joystick-Tester.shtml
Acknowledgements
This work was partly supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (NRF-2018R1C1B5046098) and the fund of research promotion program, Gyeongsang National University, 2018 and Mokwon University 2019.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Lee, S., Seo, YH. (2020). Vision-Based Virtual Joystick Interface. In: Lee, R. (eds) Computational Science/Intelligence and Applied Informatics. CSII 2019. Studies in Computational Intelligence, vol 848. Springer, Cham. https://doi.org/10.1007/978-3-030-25225-0_5
Download citation
DOI: https://doi.org/10.1007/978-3-030-25225-0_5
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-25224-3
Online ISBN: 978-3-030-25225-0
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)