Skip to main content

Learning a Head-Tracking Pointing Interface

  • Conference paper
  • First Online:
Computers Helping People with Special Needs (ICCHP-AAATE 2022)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 13341))

Abstract

For people with poor upper limb mobility or control, interaction with a computer may be facilitated by adaptive and alternative interfaces. Visual head tracking has proven to be a viable pointing interface, which can be used when use of the mouse or trackpad is challenging. We are interested in new mechanisms to map the user’s head motion to a pointer location on the screen. Towards this goal, we collected a data set of videos of participants as they were moving their head while following the motion of a marker on the screen. This data set could be used to training a machine learning system for pointing interface. We believe that by learning on real examples, this system may provide a more natural and satisfactory interface than current systems based on pre-defined algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 79.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Bates, R., Istance, H.O.: Why are eye mice unpopular? A detailed comparison of head and eye controlled assistive technology pointing devices. Univers. Access Inf. Soc. 2(3), 280–290 (2003)

    Article  Google Scholar 

  2. Betke, M., Gips, J., Fleming, P.: The camera mouse: visual tracking of body features to provide computer access for people with severe disabilities. IEEE Trans. Neural Syst. Rehabil. Eng. 10(1), 1–10 (2002)

    Article  Google Scholar 

  3. of Boston College, T.: Cameramouse (2018). http://www.cameramouse.org/. Accessed 31 Jan 2022

  4. Cicek, M., Dave, A., Feng, W., Huang, M.X., Haines, J.K., Nichols, J.: Designing and evaluating head-based pointing on smartphones for people with motor impairments. In: The 22nd International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS 2020 (2020)

    Google Scholar 

  5. Cicek, M., Xie, J., Wang, Q., Piramuthu, R.: Mobile head tracking for ecommerce and beyond. Electron. Imaging 2020(3), 303–1 (2020)

    Article  Google Scholar 

  6. Corporation, O.I.: Headmouse nano (2017). http://www.orin.com/access/headmouse/. Accessed 31 Jan 2022

  7. Deng, J., Guo, J., Zhou, Y., Yu, J., Kotsia, I., Zafeiriou, S.: Retinaface: single-stage dense face localisation in the wild. CoRR (2019). http://arxiv.org/abs/1905.00641

  8. Fanelli, G., Gall, J., Van Gool, L.: Real time head pose estimation with random regression forests. In: CVPR 2011, pp. 617–624. IEEE (2011)

    Google Scholar 

  9. Gizatdinova, Y., Špakov, O., Surakka, V.: Comparison of video-based pointing and selection techniques for hands-free text entry. In: Proceedings of the ACM International Working Conference on Advanced Visual Interfaces, pp. 132–139 (2012)

    Google Scholar 

  10. Glassouse: Glassouse assistive device (2018). http://glassouse.com/. Accessed 17 July 2018

  11. Guo, X., et al.: PFLD: a practical facial landmark detector. CoRR (2019). http://arxiv.org/abs/1902.10859

  12. Inc., A.: Use switch control to navigate your iphone, ipad, or ipod touch (2018). https://support.apple.com/en-us/ht201370. Accessed 15 July 2018

  13. Kytö, M., Ens, B., Piumsomboon, T., Lee, G.A., Billinghurst, M.: Pinpointing: precise head-and eye-based target selection for augmented reality. In: Proceedings of the 2018 ACM CHI Conference on Human Factors in Computing Systems, p. 81 (2018)

    Google Scholar 

  14. Li, J., et al.: DSFD: dual shot face detector. CoRR (2018). http://arxiv.org/abs/1810.10220

  15. LLC, P.D.: Smylemouse (2016). https://smylemouse.com/. Accessed 31 Jan 2022

  16. Majaranta, P.: Gaze Interaction and Applications of Eye Tracking: Advances in Assistive Technologies: Advances in Assistive Technologies. IGI Global (2011). https://doi.org/10.4018/978-1-61350-098-9

  17. Mauri, C.: Enable viacam (2017). http://eviacam.crea-si.com/index.php. Accessed 31 Jan 2022

  18. Mauri, C.: Eva facial mouse (2018). https://github.com/cmauri/eva_facial_mouse. Accessed 31 Jan 2022

  19. MyGaze: Mygaze assistive (2018). http://www.mygaze.com/products/mygaze-assistive/. Accessed 16 July 2018

  20. Oy, Q.: Quha zono (2018). http://www.quha.com/products-2/zono/. Accessed 15 July 2018

  21. Polacek, O., Grill, T., Tscheligi, M.: Nosetapping: what else can you do with your nose? In: Proceedings of the 12th ACM International Conference on Mobile and Ubiquitous Multimedia (2013)

    Google Scholar 

  22. Ruiz, N., Chong, E., Rehg, J.M.: Fine-grained head pose estimation without keypoints. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 2074–2083 (2018)

    Google Scholar 

  23. Tang, X., Du, D.K., He, Z., Liu, J.: Pyramidbox: a context-assisted single shot face detector. CoRR (2018). http://arxiv.org/abs/1803.07737

  24. Turturici, M., Fanucci, L.: Inertial human interface device for smartphone and tablet dedicated to people with motor disability. In: Volume 33: Assistive Technology: From Research to Practice. Assistive Technology Research Series (2013). https://doi.org/10.3233/978-1-61499-304-9-494

  25. Viola, P., Jones, M.: Rapid object detection using a boosted cascade of simple features. In: Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2001, vol. 1, pp. I-I. IEEE (2001)

    Google Scholar 

  26. Wang, X., Bo, L., Li, F.: Adaptive wing loss for robust face alignment via heatmap regression. CoRR (2019). http://arxiv.org/abs/1904.07399

  27. Wu, W., Qian, C., Yang, S., Wang, Q., Cai, Y., Zhou, Q.: Look at boundary: a boundary-aware face alignment algorithm. CoRR (2018). http://arxiv.org/abs/1805.10483

  28. Yang, T.Y., Chen, Y.T., Lin, Y.Y., Chuang, Y.Y.: Fsa-net: learning fine-grained structure aggregation for head pose estimation from a single image. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1087–1096 (2019)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Roberto Manduchi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Cicek, M., Manduchi, R. (2022). Learning a Head-Tracking Pointing Interface. In: Miesenberger, K., Kouroupetroglou, G., Mavrou, K., Manduchi, R., Covarrubias Rodriguez, M., Penáz, P. (eds) Computers Helping People with Special Needs. ICCHP-AAATE 2022. Lecture Notes in Computer Science, vol 13341. Springer, Cham. https://doi.org/10.1007/978-3-031-08648-9_46

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-08648-9_46

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-08647-2

  • Online ISBN: 978-3-031-08648-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics