Abstract
This paper describes the development of a reactive behavioral response framework for the tabletop robot Haru. The framework enables the robot to react to external stimuli through a repertoire of expressive routines. The behavioral response framework is inspired by the simple reactive behaviors of organisms (e.g. reflexes) based on a bottom-up attention model. First, a participatory study for behavior elicitation was conducted. We explored the possible expressive behaviors of the robot and the possible stimuli trigger. These stimuli-response (S-R) pairs are designed befitting the robot’s characteristics. Then, we developed a perception and a reactive behavior module that automatically translates any perceived stimulus into expressive behavioral responses. We evaluated the proposed S-R framework using Haru in an interaction setting and our results show an increase in human attention activity indicative of its positive impact to conveying the robot’s sense of agency.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
References
Admoni, H., Scassellati, B.: Social eye gaze in human-robot interaction: a review. J. Hum. Rob. Interact. 6(1), 25–63 (2017)
Belpaeme, T., Kennedy, J., Ramachandran, A., Scassellati, B., Tanaka, F.: Social robots for education: a review. Sci. Rob. 3(21) (2018)
Braitenberg, V.: Vehicles: Experiments in Synthetic Psychology. MIT Press, Cambridge (1986)
Brooks, R.A.: Elephants don’t play chess. Rob. Auton. Syst. 6(1–2), 3–15 (1990)
Cerf, M., Harel, J., Einhäuser, W., Koch, C.: Predicting human gaze using low-level saliency combined with face detection. Adv. Neural Inf. Process. Syst. 20, 1–7 (2008)
Gomez, R., Szapiro, D., Galindo, K., Nakamura, K.: Haru: hardware design of an experimental tabletop robot assistant. In: Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction, pp. 233–240 (2018)
Gomez, R., Szapiro, D., Merino, L., Nakamura, K.: A holistic approach in designing tabletop robot’s expressivity. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 1970–1976. IEEE (2020)
Kennedy, J., Baxter, P., Belpaeme, T.: The robot who tried too hard: social behaviour of a robot tutor can negatively affect child learning. In: 2015 10th ACM/IEEE International Conference on Human-Robot Interaction (HRI), pp. 67–74. IEEE (2015)
Lee, D.H., Anderson, A.K.: Reading what the mind thinks from how the eye sees. Psychol. Sci. 28(4), 494–503 (2017)
Marat, S., Phuoc, T.H., Granjon, L., Guyader, N., Pellerin, D., Guérin-Dugué, A.: Modelling spatio-temporal saliency to predict gaze direction for short videos. Int. J. Comput. Vis. 82(3), 231 (2009)
Mazzola, C., Aroyo, A.M., Rea, F., Sciutti, A.: Interacting with a social robot affects visual perception of space. In: Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction, pp. 549–557 (2020)
Mital, P.K., Smith, T.J., Hill, R.L., Henderson, J.M.: Clustering of gaze during dynamic scene viewing is predicted by motion. Cogn. Comput. 3(1), 5–24 (2011)
Nakamura, K., Gomez, R.: Improving separation of overlapped speech for meeting conversations using uncalibrated microphone array. In: 2017 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), pp. 55–62. IEEE (2017)
Peters, R.J., Itti, L.: Applying computational tools to predict gaze direction in interactive visual environments. ACM Trans. Appl. Percept. (TAP) 5(2), 1–19 (2008)
Shinn-Cunningham, B.G.: Object-based auditory and visual attention. Trends Cogn. Sci. 12(5), 182–186 (2008)
Shotton, J., et al.: Real-time human pose recognition in parts from single depth images. In: CVPR 2011, pp. 1297–1304. IEEE (2011)
Treisman, A.M., Gelade, G.: A feature-integration theory of attention. Cogn. Psychol. 12(1), 97–136 (1980)
Vasylkiv, Y., et al.: An exploration of simple reactive responses for conveying aliveness using the Haru robot. In: Wagner, A.R. (ed.) ICSR 2020. LNCS (LNAI), vol. 12483, pp. 108–119. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-62056-1_10
Walter, W.G.: An electro-mechanical «animal». dialectica, pp. 206–213 (1950)
Xu, T., Zhang, H., Yu, C.: See you see me: the role of eye contact in multimodal human-robot interaction. ACM Trans. Interact. Intell. Syst.(TiiS) 6(1), 1–22 (2016)
Vasylkiv, Y., et al.: Automating behavior selection for affective telepresence robot. In: International Conference on Robotics and Automation, ICRA. IEEE (2021)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Gomez, R. et al. (2021). Developing a Robot’s Empathetic Reactive Response Inspired by a Bottom-Up Attention Model. In: Li, H., et al. Social Robotics. ICSR 2021. Lecture Notes in Computer Science(), vol 13086. Springer, Cham. https://doi.org/10.1007/978-3-030-90525-5_8
Download citation
DOI: https://doi.org/10.1007/978-3-030-90525-5_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-90524-8
Online ISBN: 978-3-030-90525-5
eBook Packages: Computer ScienceComputer Science (R0)