Skip to main content

Eye-Interface System Using Convolutional Neural Networks for People with Physical Disabilities

  • Conference paper
  • First Online:
4th EAI International Conference on Robotic Sensor Networks

Abstract

In recent years, lack of effective communication method is one of the important issues that the people with physical disabilities, such as ALS (amyotrophic lateral sclerosis) face in social life. In this research we constructed an eye-gaze input system based on an image processing method using a PC and a web camera that are inexpensive and easy to install under natural light environment without the risk of illness due to the use of infrared radiation. In our proposed system, convolutional neural networks (CNN) is applied to improve the accuracy for practical use. The CNN in this study estimates the gaze position on the monitor screen from the image acquired from the web camera, and aims to obtain higher accuracy than the conventional system by learning for specific individuals.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Kuno, Y., Yagi, T., Fujii, I., Koga, K., Uchikawa, Y.: Development of eye-gaze input interface using EOG. J. Inf. Process. 39(5), 1455–1462 (1998)

    Google Scholar 

  2. Itoh, K., Sudoh, Y., Ifukube, T.: Eye gaze communication system for people with severe physical disabilities. IEICE Trans. Inf. Syst. Pt.1 J83-D-1(5), 495–503 (2000)

    Google Scholar 

  3. Ochiai, T., Ishimatsu, T., Takami, O., Matsui, R.: Data input device for physically handicapped people operated by eye movements. Trans. Jpn. Soc. Mech. Eng. C 63(609), 1546–1550 (1997)

    Article  Google Scholar 

  4. Kojima, S., Ukita, N., Hagita, N., Yang, M.: Simultaneous temporal optimization of pedestrians’ gaze orientation and gaze point in a scene. IPSJ SIG Tech. Rep. 2015-CG-161(21), 1–8 (2015)

    Google Scholar 

Download references

Acknowledgements

This work was supported by JSPS KAKENHI Grant-in-Aid for Scientific Research (B): Grant Number JP19H04228

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shenglin Mu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Kubo, K., Shibata, S., Karita, T., Yamamoto, T., Mu, S. (2022). Eye-Interface System Using Convolutional Neural Networks for People with Physical Disabilities. In: Mu, S., Yujie, L., Lu, H. (eds) 4th EAI International Conference on Robotic Sensor Networks. EAI/Springer Innovations in Communication and Computing. Springer, Cham. https://doi.org/10.1007/978-3-030-70451-3_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-70451-3_7

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-70450-6

  • Online ISBN: 978-3-030-70451-3

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics