Abstract
As a new type of human-robot interaction (HRI), hand gesture has many advantages such as natural operation, rich expression and not subject to environ-mental constraints. So it is very suitable for space human-robot interaction tasks in special and harsh environment. Considering that static hand gesture is one of the main gesture expressions in human-computer interaction, so a parallel convolution neural networks (CNNs) is designed to improve the accuracy of static hand gesture recognition in the conditions of complex background and changing illumination. In addition, the method is applied to the operation of space human-robot system with hand gesture control. Various space HRI hand gestures from different subjects are evaluated and tested, and experimental results demonstrate that the proposed method outperforms the single-channel CNN methods and other popular methods with a higher accuracy.
References
Fong, T., Nourbakhsh, I.: Interaction challenges in human-robot space exploration. J. Interact. 12(2), 42–45 (2005)
Liu, J.G., Luo, Y.F., Ju, Z.J.: An interactive astronaut-robot system with gesture control. J. Comput. Intell. Neurosci. 2016 (2016)
Wolf, M.T., Assad, C., Stoica, A.: Decoding static and dynamic arm and hand gestures from the JPL BioSleeve. In: Proceedings of the IEEE Aerospace Conference (2013)
Pisharady, P.K., Saerbeck, M.: Recent methods and databases in vision-based hand gesture recognition: a review. J. Comput. Vis. Image Underst. 141, 152–165 (2015)
Hasan, H., Abdul-Kareem, S.: Human-computer interaction using vision-based hand gesture recognition systems: a survey. J. Neural Comput. Appl. 25(2), 251–261 (2014)
Yamashita, T., Watasue, T.: Hand posture recognition based on bottom-up structured deep CNN with curriculum learning. In: Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP) (2014)
Molchanov, P., Mello, S.D., Kim, K., Kautz, J.: Hand gesture recognition with 3D convolutional neural networks. In: Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR 2015 (2015)
Nagi, J., Ducatelle, F., Caro, G.D., et al.: Max-pooling convolutional neural networks for vision-based hand gesture recognition. In: Proceedings of the 2011 IEEE International Conference on Signal and Image Processing Applications (ICSIPA) (2011)
Lin, H.I., Hsu, M.H., Chen, W.K.: Human hand gesture recognition using a convolution neural network. In: Proceedings of the 2014 IEEE International Conference on Automation Science and Engineering (CASE) (2014)
John, V., Boyali, A., Mita, S, et al.: Deep learning-based fast hand gesture recognition using representative frame. In: Proceedings of the 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA) (2016)
Ju, Z.J., Gao, D.X., Cao, J.T., Liu, H.H.: A novel approach to extract hand gesture feature in depth images. J. Multimedia Tools Appl. 75(19), 11929–11943 (2015)
Ju, Z.J., Wang, Y.H., Zeng, W, Cai, H.B., Liu, H.H.: An integrative framework for human hand gesture segmentation in RGB-D data. IEEE Syst. J. (in Press). doi:10.1109/JSYST.2015.2468231
Gattupalli, S, Ghaderi, A., Athitsos, V.: Evaluation of deep learning based pose estimation for sign language recognition. In: Proceedings of the 9th ACM International Conference on PErvasive Technologies Related to Assistive Environments (2016)
Ciresan, D., Meier, U., Schmidhuber, J.: Multi-column deep neural networks for image classification. In: Proceedings of the CVPR, pp. 3642–3649 (2012)
Krizhevsky, A., Sutskever I., Hinton, G.: ImageNet classification with deep convolutional neural networks. J. Adv. Neural Inf. Process. Syst. 25 (2012)
LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. J. Nat. 521, 436–444 (2015)
Pugeault, N., Bowden, R.: Spelling it out: real-time ASL fingerspelling recognition. In: Proceedings of the IEEE International Conference on Computer Vision Workshops (ICCV Workshops), pp. 1114–1119 (2011)
Odriguez, K.O., Chavez, G.C.: Finger spelling recognition from RGB-D information using kernel descriptor. In: Proceedings of the 26th SIB-GRAPI-Conference on Graphics, Patterns and Images (SIBGRAPI), pp. 1–7 (2013)
Acknowledgments
This work is supported by Research Fund of China Manned Space Engineering (050102), the Key Research Program of the Chinese Academy of Sciences (Y4A3210301), the National Science Foundation of China (51175494, 61128008, and 51575412), and the State Key Laboratory of Robotics Foundation.
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Gao, Q., Liu, J., Ju, Z., Li, Y., Zhang, T., Zhang, L. (2017). Static Hand Gesture Recognition with Parallel CNNs for Space Human-Robot Interaction. In: Huang, Y., Wu, H., Liu, H., Yin, Z. (eds) Intelligent Robotics and Applications. ICIRA 2017. Lecture Notes in Computer Science(), vol 10462. Springer, Cham. https://doi.org/10.1007/978-3-319-65289-4_44
Download citation
DOI: https://doi.org/10.1007/978-3-319-65289-4_44
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-65288-7
Online ISBN: 978-3-319-65289-4
eBook Packages: Computer ScienceComputer Science (R0)