Abstract
Gaze has a major role in social interaction. As a deictic reference, gaze aims at attracting visual attention of a communication partner to a referred entity in the environment. Gaze direction in natural faces is a well-investigated domain of research at behavioral and neurophysiological levels. However, our knowledge about deictic role of gaze in Human Robot Interaction is limited. The present study focuses on a comparative analysis of the deictic role of gaze direction in alternative face morphologies. We report an experimental study that investigated deictic gaze in a virtual reality environment. Human participants identified object locations by utilizing deictic gaze cues provided by avatar faces, as well as natural human faces. Our findings reveal a facilitating role in the accuracy of objection detection in favor of gaze embedded in natural faces compared to gaze embedded in synthetic avatar face morphologies.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Perrett, D., Smith, P., Potter, D., et al.: Visual cells in the temporal cortex sensitive to face view and gaze direction. In: Proceedings of the Royal Society of London, Series B, Biological Sciences, vol. 223, no. 1232, pp. 293–317. Royal Society (1985)
Stiefelhagen, R., Zhu, J.: Head orientation and gaze direction in meetings. In: CHI 2002 Extended Abstracts on Human Factors in Computing Systems (CHI EA 2002), pp. 858–859. Association for Computing Machinery, New York City (2002)
Katzenmaier, M., Stiefelhagen, B., Schultz, T.: Identifying the addressee in human-human-robot interactions based on head pose and speech. In: ICMI 2004 - Sixth International Conference on Multimodal Interfaces, pp. 144–151. Association for Computing Machinery, New York City (2004)
Tan, Z., Thomsen, N.B., Duan, X., et al.: iSocioBot: a multimodal interactive social robot. Int. J. Soc. Robot. 10(1), 5–19 (2018)
Galantucci, B.: An experimental study of the emergence of human communication systems. Cogn. Sci. 29(5), 737–767 (2005)
Sebanz, N., Bekkering, H., Knoblich, G.: Joint action: bodies and minds moving together. Trends Cogn. Sci. 10(2), 70–76 (2006)
Haddadin, S., Croft, E.: Erratum to: Physical human–robot interaction. In: Siciliano, B., Khatib, O. (eds.) Springer Handbook of Robotics, p. E1. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-32552-1_81
Fang, R., Doering, M., Chai, J.Y.: Embodied collaborative referring expression generation in situated human-robot interaction. In: Proceedings of the 10th Annual ACM/IEEE International Conference on Human-Robot Interaction, pp. 271–278. Association for Computing Machinery, New York City (2015)
Lemaignan, S., Warnier, M., Sisbot, E.A., et al.: Artificial cognition for social human–robot interaction: an implementation. Artif. Intell. 247(1), 45–69 (2017)
Piwek, P.: Salience in the generation of multimodal referring acts. In: Proceedings of the 2009 International Conference on Multimodal Interfaces, pp. 207–210 (2009)
Admoni, H., Scassellati, B.: Social eye gaze in human-robot Interaction: a review. J. Hum.-Robot Interact. 6(1), 25–63 (2017)
Ruhland, K., Peters, C.E., Andrist, S., et al.: A review of eye gaze in virtual agents, social robotics and HCI: behaviour generation, user interaction and perception. Comput. Graph. Forum 34(6), 299–326 (2015)
Yücel, Z., Salah, A.A., Meriçli, Ç., et al.: Joint attention by gaze interpolation and saliency. IEEE Trans. Cybern. 43(3), 829–842 (2013)
Breazeal, C., Dautenhahn, K., Kanda, T.: Social robots. In: Siciliano, B., Khatib, O. (eds.), Springer Handbook of Robotics. 2nd edn. pp. 1935–1961. Springer Handbooks, Heidelberg (2016).https://doi.org/10.1007/978-3-319-32552-1
Devault, D., Kariaeva, N., Kothari, A., et al.: An information-state approach to collaborative reference. In: Proceedings of the ACL Interactive Poster and Demonstration Sessions, pp. 1–4 (2005)
Mutlu, B., Shiwa, T., Kanda, T., et al.: Footing in human-robot conversations: how robots might shape participant roles using gaze cues. Hum. Factors 2(1), 61–68 (2009)
Shannon, C.E.: A mathematical theory of communication. Bell Syst. Tech. J. 1(1), 379–423 (1948)
Holmqvist, K., Andersson, R.: Paradigms - visual search. In: Eye Tracking: A Comprehensive Guide to Methods and Measures, pp. 409–413. Oxford University Press, Oxford (2011)
Acknowledgments
This project has been supported by TÜBİTAK 117E021 “A gaze-mediated framework for multimodal Human Robot Interaction”.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Yılmaz, E., Fal, M., Acartürk, C. (2020). The Role of Gaze as a Deictic Cue in Human Robot Interaction. In: Schmorrow, D., Fidopiastis, C. (eds) Augmented Cognition. Human Cognition and Behavior. HCII 2020. Lecture Notes in Computer Science(), vol 12197. Springer, Cham. https://doi.org/10.1007/978-3-030-50439-7_32
Download citation
DOI: https://doi.org/10.1007/978-3-030-50439-7_32
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-50438-0
Online ISBN: 978-3-030-50439-7
eBook Packages: Computer ScienceComputer Science (R0)