Skip to main content

Recognition and Localisation of Pointing Gestures Using a RGB-D Camera

  • Conference paper
  • First Online:
HCI International 2020 - Posters (HCII 2020)

Abstract

Non-verbal communication is part of our regular conversation, and multiple gestures are used to exchange information. Among those gestures, pointing is the most important one. If such gestures cannot be perceived by other team members, e.g. by blind and visually impaired people (BVIP), they lack important information and can hardly participate in a lively workflow. Thus, this paper describes a system for detecting such pointing gestures to provide input for suitable output modalities to BVIP. Our system employs an RGB-D camera to recognize the pointing gestures performed by the users. The system also locates the target of pointing e.g. on a common workspace. We evaluated the system by conducting a user study with 26 users. The results show that the system has a success rate of 89.59 and 79.92 % for a \(2 \times 3\) matrix using the left and right arm respectively, and 73.57 and 68.99% for \(3 \times 4\) matrix using the left and right arm respectively.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://www.ros.org/.

References

  1. Brannigan, C.R., Humphries, D.A.: Human non-verbal behavior, a means of communication. Etholog. Stud. Child Behav., 37–64 (1972)

    Google Scholar 

  2. Carraro, M., Munaro, M., Burke, J., Menegatti, E.: Real-time marker-less multi-person 3D pose estimation in rgb-depth camera networks. In: Strand, M., Dillmann, R., Menegatti, E., Ghidoni, S. (eds.) IAS 2018. AISC, vol. 867, pp. 534–545. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-01370-7_42

    Chapter  Google Scholar 

  3. Dhingra, N., Kunz, A.: Res3atn-deep 3D residual attention network for hand gesture recognition in videos. In: International Conference on 3D Vision (3DV 2019), pp. 491–501. IEEE (2019)

    Google Scholar 

  4. Günther, S., et al.: Mapvi: meeting accessibility for persons with visual impairments. In: Proceedings of the 12th ACM International Conference on PErvasive Technologies Related to Assistive Environments, pp. 343–352. ACM (2019)

    Google Scholar 

  5. Huang, Y., Liu, X., Zhang, X., Jin, L.: A pointing gesture based egocentric interaction system: dataset, approach and application. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 16–23 (2016)

    Google Scholar 

  6. LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436 (2015)

    Article  Google Scholar 

  7. Munaro, M., Basso, F., Menegatti, E.: Openptrack: open source multi-camera calibration and people tracking for RGB-D camera networks. Robot. Auton. Syst. 75, 525–538 (2016)

    Article  Google Scholar 

  8. Neverova, N., Wolf, C., Taylor, G.W., Nebout, F.: Multi-scale deep learning for gesture detection and localization. In: Agapito, L., Bronstein, M.M., Rother, C. (eds.) ECCV 2014. LNCS, vol. 8925, pp. 474–490. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-16178-5_33

    Chapter  Google Scholar 

  9. Park, C.B., Lee, S.W.: Real-time 3D pointing gesture recognition for mobile robots with cascade HMM and particle filter. Image Vis. Comput. 29(1), 51–63 (2011)

    Article  Google Scholar 

  10. Quam, D.L.: Gesture recognition with a dataglove. In: IEEE Conference on Aerospace and Electronics, pp. 755–760. IEEE (1990)

    Google Scholar 

  11. Rautaray, S.S., Agrawal, A.: Vision based hand gesture recognition for human computer interaction: a survey. Artif. Intell. Rev. 43(1), 1–54 (2012). https://doi.org/10.1007/s10462-012-9356-9

    Article  Google Scholar 

  12. Wilson, A.D., Bobick, A.F.: Parametric hidden markov models for gesture recognition. IEEE Trans. Pattern Anal. Mach. Intell. 21(9), 884–900 (1999)

    Article  Google Scholar 

Download references

Acknowledgements

This work has been supported by the Swiss National Science Foundation (SNF) under the grant no. 200021E 177542/1. It is part of a joint project between TU Darmstadt, ETH Zurich, and JKU Linz with the respective funding organizations DFG (German Research Foundation), SNF (Swiss National Science Foundation) and FWF (Austrian Science Fund).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Naina Dhingra .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Dhingra, N., Valli, E., Kunz, A. (2020). Recognition and Localisation of Pointing Gestures Using a RGB-D Camera. In: Stephanidis, C., Antona, M. (eds) HCI International 2020 - Posters. HCII 2020. Communications in Computer and Information Science, vol 1224. Springer, Cham. https://doi.org/10.1007/978-3-030-50726-8_27

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-50726-8_27

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-50725-1

  • Online ISBN: 978-3-030-50726-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics