Skip to main content

Hand Location Classification from 3D Signing Virtual Avatars Using Neural Networks

  • Conference paper
Computers Helping People with Special Needs (ICCHP 2014)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 8548))

Included in the following conference series:

Abstract

3D sign language data is actively being generated and exchanged. Sign language recognition from 3D data is then a promising research axis aiming to build new understanding and efficient indexing of this type of content. Model-based recognition strategies are commonly based on recognizing sign language features separately. Those features are: the handshape, the hand position, the orientation and movement. In this paper, we propose a novel approach for hand position classification in the space. The approach is based on a two-layer feed-forward network and generates classifications which are very close to human perception. Evaluations have been made by 10 PhD students and 2 sign language experts. The evaluation of the results shows the superiority of our approach compared with classic methods based on the calculation of the distance between the hand and the face as well as the method of K nearest neighbors. In fact, the misclassification average of our methods was the lowest with 4.58%.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Huenerfauth, M., et al.: Evaluating American Sign Language generation through the participation of native ASL signers. In: ACM SIGACCESS Conference on Computers and Accessibilit. ACM, New York (2007)

    Google Scholar 

  2. Buttussi, F., Chittaro, L., Nadalutti, D.: H-Animator: A Visual Tool for Modeling, Reuse and Sharing of X3D Humanoid Animations. In: International Conference on 3D Web Technology, New York (2006)

    Google Scholar 

  3. Jemni, M., Elghoul, O.: A System to Make Signs Using Collaborative Approach. In: Miesenberger, K., Klaus, J., Zagler, W.L., Karshmer, A.I. (eds.) ICCHP 2008. LNCS, vol. 5105, pp. 670–677. Springer, Heidelberg (2008)

    Chapter  Google Scholar 

  4. Zwiterslood, I., Verlinden, M., Ros, J., van der Schoot, S.: Synthetic Signing for the Deaf: eSIGN. In: Conference and Workshop on Assistive Technologies for Vision and Hearing Impairment, Granada (2004)

    Google Scholar 

  5. Papadogiorgaki, M., Grammalidis, N., Sarris, N., Strintzis, M.G.: Synthesis of Virtual Reality Animations from SWML using MPEG-4 Body Animation Parameters. In: Sign Processing Workshop, Lisbon (2004)

    Google Scholar 

  6. Vogler, C., Metaxas, D.: ASL recognition based on a coupling between HMMs and 3D motion analysis. In: Sixth International Conference on Computer Vision (1998)

    Google Scholar 

  7. Liddell, S., Johnson, R.: American Sign Language: The Phonological Base. Sign Language Studies (164)

    Google Scholar 

  8. Huenerfauth, M.: Improving Spatial Reference in American Sign Language Animation through Data Collection from Native ASL Signers. In: Stephanidis, C. (ed.) Universal Access in HCI, Part III, HCII 2009. LNCS, vol. 5616, pp. 530–539. Springer, Heidelberg (2009)

    Chapter  Google Scholar 

  9. Perniss, P.M.: Achieving spatial coherence in German Sign Language narratives: The use of classifiers and perspective. Lingua 117(17), 1315–1338 (2007)

    Article  Google Scholar 

  10. Liang, R., Ouhyoung, M.: A real-time continuous gesture recognition system for sign language, In: International Conference on Automatic Face and Gesture Recognition, Nara, Japan (1998)

    Google Scholar 

  11. Stokoe, W.C.: Sign Language Structure: An Outline of the Visual Communication Systems of the American Deaf. Studies in Linguistics: Occasional Papers (18) (1960)

    Google Scholar 

  12. Han, J., Awad, G., Sutherland, A.: Modelling and segmenting subunits for sign language recognition based on hand motion analysis. Pattern Recognition Letters 30(16), 623–633 (2009)

    Article  Google Scholar 

  13. Ding, L., Martinez, A.M.: Modelling and recognition of the linguistic components in American Sign Language. Image and Vision Computing 27(112), 1826–1844 (2009)

    Article  Google Scholar 

  14. Oz, C., Leu, M.: American Sign Language word recognition with a sensory glove using artificial neural networks. Engineering Applications of Artificial Intelligence 24(17), 1204–1213 (2011)

    Article  Google Scholar 

  15. Braffot, A., Dalle, P.: Sign language applications: preliminary modeling. Universal Access in the Information Society 4 (2008)

    Google Scholar 

  16. Bangham, J., Cox, S., Elliott, R., Glauert, J., Marshall, I., Rankov, S., Wells, M.: Virtual Signing: Capture, Animation, Storage and Transmission - An Overview of the ViSiCAST Project. In: IEEE Seminar on Speech and Language Processing for Disabled and Elderly People (2000)

    Google Scholar 

  17. Jaballah, K., Jemni, M.: Toward Automatic Sign Language Recognition from Web3D Based Scenes. In: Miesenberger, K., Klaus, J., Zagler, W., Karshmer, A. (eds.) ICCHP 2010, Part II. LNCS, vol. 6180, pp. 205–212. Springer, Heidelberg (2010)

    Chapter  Google Scholar 

  18. Masters, T.: Advanced algorithms for neural networks: a C++ sourcebook. John Wiley & Sons, NY (1995)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this paper

Cite this paper

Jaballah, K., Jemni, M. (2014). Hand Location Classification from 3D Signing Virtual Avatars Using Neural Networks. In: Miesenberger, K., Fels, D., Archambault, D., Peňáz, P., Zagler, W. (eds) Computers Helping People with Special Needs. ICCHP 2014. Lecture Notes in Computer Science, vol 8548. Springer, Cham. https://doi.org/10.1007/978-3-319-08599-9_66

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-08599-9_66

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-08598-2

  • Online ISBN: 978-3-319-08599-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics