Automatic Recognition of Sign Language Images
The objective of the research presented in this chapter is to enable communication between people with hearing impairment and those with visual impairment. Computer recognition of sign language snapshots is one of the most challenging research problems in this area. This chapter presents an efficient and fast algorithm for identification of the number of fingers opened in a gesture representing an alphabet of the American Sign Language. Finger detection is accomplished based on the concept of boundary tracing and finger tip detection. A significant feature of the solution is that it does not require the hand to be perfectly aligned to the camera or use any special markers or input gloves.
KeywordsBoundary tracing Computer access for disabled Finger detection Image processing Sign language recognition
The authors would like to acknowledge the contribution of Bharadvaj J., Ganesh S., Ravindra K., Vinod D. towards the development of one part of the solution presented in this chapter.
- 2.Shimada, N. et al. (1998). Hand gesture estimation and model refinement using monocular camera – Ambiguity limitation by inequality constraints. In Proceedings of the 3rd Conference on Face and Gesture Recognition, pp. 268–273.Google Scholar
- 3.Sonka, M., Hlavac, V., & Boyle, R. (1999). Image processing, analysis, and machine vision. Pacific Grove, CA: Brooks/Cole Publishing Company.Google Scholar
- 4.Starner, T., & Pentland, A. (21–23 Nov 1995) Real-time American sign language recognition from video using hidden Markov models. In Proceedings of International Symposium on Computer Vision, pp. 265–270.Google Scholar
- 6.Fillbrandt, H., Akyol, S., & Kraiss, K.F. (Oct 2003). Extraction of 3D hand shape and posture from image sequences for sign language recognition. IEEE International Workshop on Analysis and Modeling of Faces and Gestures, 17, 181–186.Google Scholar
- 8.Imagawa, K., Lu, S., & Igi, S. (14–16 Apr 1998) Color-based hands tracking system for sign language recognition. In Proceedings of the Third IEEE International Conference on Automatic Face and Gesture Recognition, pp. 462–467.Google Scholar
- 10.Kang, S.K., Nam, M.Y., & Rhee, P.K. (Aug 2008) Color based hand and finger detection technology for user interaction. International Conference on Convergence and Hybrid Information Technology.Google Scholar
- 11.Shirai, Y., Tanibata, N., & Shimada, N. (2002). Extraction of hand features for recognition of sign language words. VI’2002, Computer-Controlled Mechanical Systems. Graduate School of Engineering, Osaka University.Google Scholar
- 12.Hamada, Y., Shimada, N., & Shirai, Y. (May 2004) Hand shape estimation under complex backgrounds for sign language recognition. In Proceedings of the 6th International Conference on Automatic Face and Gesture Recognition, pp. 589–594.Google Scholar
- 13.Ravikiran, J., Mahesh, K., Mahishi, S., Dheeraj, R., Sudheender, S., & Pujari, N.V. (2009). Finger detection for sign language recognition, ICCS2009. In Proceedings of the International Conference on Computer Science, International Multi-Conference of Engineers and Computer Scientists – IMECS 2009, March 2009, Hong Kong.Google Scholar