Abstract
Humans communicate on three levels: words, paralanguage, and nonverbal. While conversational agents focus mainly on the interpretation of words that are being spoken, recently the focus has also shifted to how we say those words with our tone, pace, and intonation. Nonverbal communication, including facial expression, eye contact, posture, and proximity, has been largely ignored in human-agent interactions.
In this work, we propose to incorporate nonverbal behavior into conversations between humans and agents by displaying a human-like embodied agent on a large screen and by responding appropriately to nonverbal cues from the interlocutors. In a user study with 19 volunteers, we investigated the influence on the participants for different behaviors (mimicry, positively biased mimicry, negatively biased mimicry, random) of the embodied conversation agents. The results indicate that goal-directed behavior is perceived significantly better concerning likability, social competence, attitude, and responsiveness in comparison to random behavior. This indicates that already simple nonverbal methods of building rapport can be used to improve the perceived conversational quality with an embodied conversational agent.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
- 2.
Body Tracking SDK for Azure Kinect enables segmentation of exposed instances and both observed and estimated 3D joints and landmarks for fully articulated, uniquely identified body tracking of skeletons. (www.azure.microsoft.com/en-us/services/kinect-dk).
- 3.
Kinetic Space is an open-source tool that enables training, analysis, and recognition of individual gestures with a depth camera like Microsoft’s Kinect family [32].
References
AbuShawar, B., Atwell, E.: Alice chatbot: trials and outputs. Computación y Sistemas 19(4), 625–632 (2015)
Albrecht, I., Schröder, M., Haber, J., Seidel, H.P.: Mixed feelings: expression of non-basic emotions in a muscle-based talking head. Virt. Real. 8(4), 201–212 (2005)
Argyle, M.: Bodily Communication, 2nd edn, pp. 1–111. Routledge, London (1986)
Bickmore, T., Cassell, J.: Small talk and conversational storytelling in embodied conversational interface agents. In: AAAI Fall Symposium on Narrative Intelligence, pp. 87–92 (1999)
Bull, P.: Gesture and Posture. Pergamon Press, Oxford (1987)
Butz, M., Hepperle, D., Wölfel, M.: Influence of visual appearance of agents on presence, attractiveness, and agency in virtual reality. In: Wölfel, M., Bernhardt, J., Thiel, S. (eds.) 10th EAI International Conference, ArtsIT 2021. LNICS, vol. 422, p. 44. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-95531-1_4
Cassell, J., et al.: Embodiment in conversational interfaces: Rea. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 520–527 (1999)
Cerekovic, A., Aran, O., Gatica-Perez, D.: Rapport with virtual agents: what do human social cues and personality explain? IEEE Trans. Affect. Comput. 8(3), 382–395 (2017)
Chartrand, T.L., Bargh, J.A.: The chameleon effect: the perception-behavior link and social interaction. J. Pers. Soc. Psychol. 76(6), 893–910 (1999)
Clark, L., et al.: What makes a good conversation? Challenges in designing truly conversational agents. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1–12. Association for Computing Machinery, New York (2019)
Das, A., Datta, S., Gkioxari, G., Lee, S., Parikh, D., Batra, D.: Embodied question answering. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–10 (2018)
Duffy, K.A., Chartrand, T.L.: Mimicry: causes and consequences. Curr. Opin. Behav. Sci. 3, 112–116 (2015)
Duncan, S., Fiske, D.W.: Face to Face Interaction Research. Methods and Theory (1977)
Gong, L.: Is happy better than sad even if they are both non-adaptive? Effects of emotional expressions of talking-head interface agents. Int. J. Hum. Comput. Stud. 65(3), 183–191 (2007)
Gratch, J., et al.: Virtual Rapport. In: Gratch, J., Young, M., Aylett, R., Ballin, D., Olivier, P. (eds.) IVA 2006. LNCS (LNAI), vol. 4133, pp. 14–27. Springer, Heidelberg (2006). https://doi.org/10.1007/11821830_2
Harrigan, J.A., Oxman, T.E., Rosenthal, R.: Rapport expressed through nonverbal behavior. J. Nonverbal Behav. 9(2), 95–110 (1985)
Harrigan, J.A., Rosenthal, R.: Physicians’ head and body positions as determinants of perceived rapport. J. Appl. Soc. Psychol. 13(6), 496–509 (1983)
Heylen, D.: Challenges ahead: head movements and other social acts in conversations. Virt. Soc. Agents, 45–52 (2005)
Hoque, M.E., Courgeon, M., Martin, J.C., Mutlu, B., Picard, R.W.: MACH: my automated conversation coach. In: Proceedings of the 2013 ACM International Joint Conference on Pervasive and Ubiquitous Computing, UbiComp 2013, pp. 697–706. Association for Computing Machinery, New York (2013)
Huang, L., Morency, L.-P., Gratch, J.: Virtual Rapport 2.0. In: Vilhjálmsson, H.H., Kopp, S., Marsella, S., Thórisson, K.R. (eds.) IVA 2011. LNCS (LNAI), vol. 6895, pp. 68–79. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-23974-8_8
Lester, J., Branting, K., Mott, B.: Conversational Agents. The Practical Handbook of Internet Computing, pp. 220–240 (2004)
Lucas, G.M., et al.: Getting to know each other: the role of social dialogue in recovery from errors in social robots. In: Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction, pp. 344–351. ACM, Chicago (2018)
Mara, M., Appel, M.: Effects of lateral head tilt on user perceptions of humanoid and android robots. Comput. Hum. Behav. 44, 326–334 (2015)
Massaro, D.W., Cohen, M.M., Daniel, S., Cole, R.A.: Developing and evaluating conversational agents. In: Human Performance and Ergonomics, pp. 173–194. Elsevier (1999)
Nass, C., Moon, Y.: Machines and mindlessness: social responses to computers. J. Soc. Issues 56(1), 81–103 (2000)
Pentland, A.: Socially aware, computation and communication. Computer 38(3), 33–40 (2005)
Poggi, I., D’Errico, F., Vincze, L.: Types of nods. The polysemy of a social signal. In: Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC 2010) (2010)
Salem, B., Earle, N.: Designing a non-verbal language for expressive avatars. In: Proceedings of the Third International Conference on Collaborative Virtual Environments, CVE 2000, pp. 93–101. Association for Computing Machinery, New York (2000)
Schiaffino, S., Amandi, A.: User-interface agent interaction: personalization issues. Int. J. Hum. Comput. Stud. 60(1), 129–148 (2004)
Technologies, U.: Unity—2D 3D Game Creator & Editor—Augmented. Virtual Reality Software—Game Engine. https://unity.com/products/unity-platform
Wang, F.Y., Carley, K.M., Zeng, D., Mao, W.: Social computing: from social informatics to social intelligence. IEEE Intell. Syst. 22(2), 79–83 (2007)
Wölfel, M.: Kinetic Space - 3D Gestenerkennung für Dich und Mich. Konturen 32, 58–63 (2012)
Zakharov, E., Shysheya, A., Burkov, E., Lempitsky, V.: Few-shot adversarial learning of realistic neural talking head models. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9459–9468 (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Wölfel, M., Purps, C.F., Percifull, N. (2022). Enabling Embodied Conversational Agents to Respond to Nonverbal Behavior of the Communication Partner. In: Kurosu, M. (eds) Human-Computer Interaction. User Experience and Behavior. HCII 2022. Lecture Notes in Computer Science, vol 13304. Springer, Cham. https://doi.org/10.1007/978-3-031-05412-9_40
Download citation
DOI: https://doi.org/10.1007/978-3-031-05412-9_40
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-05411-2
Online ISBN: 978-3-031-05412-9
eBook Packages: Computer ScienceComputer Science (R0)