Skip to main content

Toward Scalability in ASL Recognition: Breaking Down Signs into Phonemes

  • Conference paper
  • First Online:
Gesture-Based Communication in Human-Computer Interaction (GW 1999)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 1739))

Included in the following conference series:

Abstract

In this paper we present a novel approach to continuous, whole-sentence ASL recognition that uses phonemes instead of whole signs as the basic units. Our approach is based on a sequential phonological model of ASL. According to this model the ASL signs can be broken into movements and holds, which are both considered phonemes. This model does away with the distinction between whole signs and epenthesis movements that we made in previous work [17]. Instead, epenthesis movements are just like the other movements that constitute the signs.

We subsequently train Hidden Markov Models (HMMs) to recognize the phonemes, instead of whole signs and epenthesis movements that we recognized previously [17]. Because the number of phonemes is limited, HMM-based training and recognition of the ASL signal becomes computationally more tractable and has the potential to lead to the recognition of large-scale vocabularies.

We experimented with a 22 word vocabulary, and we achieved similar recognition rates with phoneme-and word-based approaches. This result is very promising for scaling the task in the future.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. A. Braffort. ARGo: An architecture for sign language recognition and interpretation. In A. D. N. Edwards P. A. Harling, editor, Progress in gestural interaction. Proceedings of Gesture Workshop’ 96, pages 17–30, Berlin, New York, 1997. Springer.

    Google Scholar 

  2. D. Brentari. Sign language phonology: ASL. In J. A. Goldsmith, editor, The Handbook of Phonological Theory, Blackwell Handbooks in Linguistics, pages 615–639. Blackwell, Oxford, 1995.

    Google Scholar 

  3. G. R. Coulter, editor. Current Issues in ASL Phonology, volume 3 of Phonetics and Phonology. Academic Press, Inc., San Diego, CA, 1993.

    Google Scholar 

  4. R. Erenshteyn and P. Laskov. A multi-stage approach to ngerspelling and gesture recognition. Proceedings of the Workshop on the Integration of Gesture in Language and Speech, Wilmington, DE, USA, 1996.

    Google Scholar 

  5. S. Gibet, J. Richardson, T. Lebourque, and A. Braffort. Corpus of 3d natural movements and sign language primitives of movement. In I. Wachsmuth and M. Fröhlich, editors, Gesture and Sign Language in Human-Computer Interaction. Proceedings of Gesture Workshop’ 97, Berlin, New York, 1998. Springer.

    Google Scholar 

  6. K. Grobel and M. Assam. Isolated sign language recognition using hidden Markov models. SMC, pages 162–167, Orlando, FL, 1997.

    Google Scholar 

  7. H. Hienz, K.-F. Kraiss, and B. Bauer. Continuous sign language recognition using hidden Markov models. In Y. Tang, editor, ICMI’99, pages IV10–IV15, Hong Kong, 1999.

    Google Scholar 

  8. M. W. Kadous. Machine recognition of Auslan signs using PowerGloves: Towards large-lexicon recognition of sign language. In Proceedings of the Workshop on the Integration of Gesture in Language and Speech, pages 165–174, Wilmington, DE, USA, 1996.

    Google Scholar 

  9. R.-H. Liang and M. Ouhyoung. A real-time continuous gesture recognition system for sign language. In Proceedings of the Third International Conference on Automatic Face and Gesture Recognition, pages 558–565, Nara, Japan, 1998.

    Google Scholar 

  10. S. K. Liddell and R. E. Johnson. American Sign Language: The phonological base. Sign Language Studies, 64:195–277, 1989.

    Google Scholar 

  11. Y. Nam and K. Y. Wohn. Recognition of space-time hand-gestures using hidden Markov model. ACM Symposium on Virtual Reality Software and Technology, 1996.

    Google Scholar 

  12. L. R. Rabiner. A tutorial on Hidden Markov Models and selected applications in speech recognition. Proceedings of the IEEE, 77(2):257–286, 1989.

    Article  Google Scholar 

  13. W. Sandler. Phonological Representation of the Sign: Linearity and Nonlinearity in American Sign Language. Number 32 in Publications in Language Sciences. Foris Publications, Dordrecht, 1989.

    Google Scholar 

  14. T. Starner and A. Pentland. Visual recognition of American Sign Language using Hidden Markov Models. International Workshop on Automatic Face and Gesture Recognition, pages 189–194, Zürich, Switzerland, 1995.

    Google Scholar 

  15. W. C. Stokoe. Sign Language Structure: An Outline of the Visual Communication System of the American Deaf. Studies in Linguistics: Occasional Papers 8. Linstok Press, Silver Spring, MD, 1960. Revised 1978.

    Google Scholar 

  16. C. Vogler and D. Metaxas. Parallel hidden Markov models for American Sign Language recognition. ICCV, Kerkyra, Greece, 1999.

    Google Scholar 

  17. C. Vogler and D. Metaxas. Adapting hidden Markov models for ASL recognition by using three-dimensional computer vision methods. SMC, pages 156–161, Orlando, FL, 1997.

    Google Scholar 

  18. C. Vogler and D. Metaxas. ASL recognition based on a coupling between HMMs and 3D motion analysis. ICCV, pages 363–369, Mumbai, India, 1998.

    Google Scholar 

  19. M. B. Waldron and S. Kim. Isolated ASL sign recognition system for deaf persons. IEEE Transactions on Rehabilitation Engineering, 3(3):261–71, September 1995.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1999 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Vogler, C., Metaxas, D. (1999). Toward Scalability in ASL Recognition: Breaking Down Signs into Phonemes. In: Braffort, A., Gherbi, R., Gibet, S., Teil, D., Richardson, J. (eds) Gesture-Based Communication in Human-Computer Interaction. GW 1999. Lecture Notes in Computer Science(), vol 1739. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-46616-9_19

Download citation

  • DOI: https://doi.org/10.1007/3-540-46616-9_19

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-66935-7

  • Online ISBN: 978-3-540-46616-1

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics