Abstract
This chapter surveys some of the principal research trends in social robotics and its application to human–robot interaction (HRI). Social (or sociable) robots are designed to interact with people in a natural, interpersonal manner – often to achieve social-emotional goals in diverse applications such as education, health, quality of life, entertainment, communication, and collaboration. The long-term goal of creating social robots that are competent and capable partners for people is quite a challenging task. They will need to be able to communicate naturally with people using both verbal and nonverbal signals. They will need to engage us not only on a cognitive level, but on an emotional level as well. They will need a wide range of social-cognitive skills and a theory of other minds to understand human behavior, and to be intuitively understood by people. A deep understanding of human intelligence and behavior across multiple dimensions (i.e., cognitive, affective, physical, social, etc.) is necessary in order to design robots that can successfully play a beneficial role in the daily lives of people. This requires a multidisciplinary approach where the design of social robot technologies and methodologies are informed by robotics, artificial intelligence, psychology, neuroscience, human factors, design, anthropology, and more.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Abbreviations
- AAAI:
-
American Association for Artificial Intelligence
- ACM:
-
Association of Computing Machinery
- ACM:
-
active cord mechanism
- ASKA:
-
receptionist robot
- ATR:
-
Advanced Telecommunications Research Institute International
- FST:
-
finite-state transducer
- HAMMER:
-
hierarchical attentive multiple models for execution and recognition
- HRI:
-
human–robot interaction
- IEEE:
-
Institute of Electrical and Electronics Engineers
- NASA:
-
National Aeronautics and Space Agency
- QRIO:
-
quest for curiosity
- WABIAN:
-
Waseda bipedal humanoid
References
K. Dautenhahn: Getting to know each other – Artificial social intelligence for autonomous robots, Robot. Auton. Syst. 16, 333–356 (1995)
K. Dautenhahn: I could be you: The phenomenological dimension of social understanding, Cybern. Syst. 28, 417–453 (1997)
C. Breazeal: Designing Sociable Robots (MIT Press, Cambridge 2002)
H. Miwa, A. Takanishi, H. Takanobu: Experimental study on robot personality for humanoid head robot, Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst. (2001) pp. 1183–1188
T. Tojo, Y. Matsusaka, T. Ishii, T. Kobayashi: A conversational robot utilizing facial and body expressions, Proc. IEEE SMC2000, Vol. 2 (2000) pp. 858–863
J. Cassell, J. Sullivan, S. Prevost, E. Churchill: Embodied Conversational Agents (MIT Press, Boston 2000)
G. Hoffman, C. Breazeal: Robots that work in collaboration with people, Proc. AAAI Fall Symp. Intersec. Cognit. Sci. Robot.: From Interfaces Intell. Washington (AAAI, Menlo Park 2004)
K. Dautenhahn, A.H. Bond, L. Canamero, B. Edmonds: Socially Intelligent Agents: Creating Relationships with Computers and Robots, Multiagent Syst.Artif. Soc. Simul. Organiz., Vol. 3 (Springer, Berlin, Heidelber 2002)
W.R. Picard: Affective Computing (MIT Press, Cambridge 1997)
T. Fong, I. Nourbakshsh, K. Dautenhahn: A survey of social robots, Robot. Auton. Syst. 42, 143–166 (2003)
S. Schaal: Is imitation learning the route to humanoid robots?, Trends Cognit. Sci. 3(6), 233–242 (1999)
J. Solis, K. Chida, K. Suefuji, A. Takanishi: The development of the anthropomorphic flutist robot at Waseda University, Int. J. Humano. Robots 3(2), 1–25 (2006)
Y. Ogura, H. Aikawa, K. Shimomura, H. Kondo, A. Morishima, H. Lim, A. Takanishi: Development of a new humanoid robot WABIAN-2, Proc. IEEE Int. Conf. Robot. Autom. (2006) pp. 76–81
H. Miwa, K. Itoh, H. Takanobu, A. Takanishi: Mechanical design and motion control of emotion expression humanoid robot WE-4R, 15th CISM-IFToMM Symp.Robot Des.Dyn. Cont. ROM04–14 (2004)
F. Tanaka, J.R. Movellan, B. Fortenberry, K. Aisaka: Daily HRI evaluation at a classroom environment: Reports from dance interaction experiments, Proc. 1st Annu. Conf. Human-Robot Interact. (HRI) (Salt Lake City 2006) pp. 3–9
H. Lim, S. Hyon, S. Setiawan, S.A. Takanishi: Quasi-human biped walking, Int. J. Inform. Educ. Res. Robot. Artif. Intell. 24(2), 257–268 (2006)
H. Lim, A. Ishii, A. Takanishi: Emotion-based biped walking, Int. J. Inform. Educ. Res. Robot. Artif. Intell. 22(5), 577–586 (2004)
J. Solis, S. Isoda, K. Chida, A. Takanishi, K. Wakamatsu: Anthropomorphic flutist robot for teaching flute playing to beginner students, Proc. IEEE Int. Conf. Robot. Autom. (2004) pp. 146–151
H. Miwa, K. Itoh, M. Matsumoto, M. Zecca, H. Takanobu, S. Roccella, C.M. Carrozza, P. Dario, A. Takanishi: Effective emotional expressions with emotion expression humanoid robot WE-4RII, Proc. 2004 IEEE/RSJ Int. Conf. Intell. Robots Syst. (2004) pp. 2203–2208
S. Roccella, M.C. Carrozza, G. Cappiello, P. Dario, J. Cabibihan, M. Zecca, H. Miwa, K. Itoh, M. Matsumoto, A. Takanishi: Design, fabrication and preliminary results of a novel anthropomorphic hand for humanoid robotics: RCH-1, Proc. 2004 IEEE/RSJ Int. Conf. Intell. Robots Syst. (Sendai, 2004) pp. 266–271
H. Miwa, K. Itoh, H. Takanobu, A. Takanishi: Design and control of 9-DoFs emotion expression humanoid arm, 2004 IEEE Int. Conf. Robot. Autom. (New Orleans 2004) pp. 128–133
F. Iida, M. Tabata, F. Hara: Generating personality character in a Face Robot through interaction with human, Proc. 7th IEEE Int. Workshop Robot Human Commun. (1998) pp. 481–486
M. Shimada, T. Minato, S. Itakura, H. Ishiguro: Evaluation of Android Using Unconscious Recognition, Proc. 6th IEEE-RAS Int. Conf. Humano. Robots (2006) pp. 157–162
K. Berns, J.J. Hirth: Control of facial expressions of the humanoid robot head ROMAN, Proc. 2006 IEEE/RSJ Int. Conf. Intell. Robots Syst. (Bejing 2006) pp. 3119–3124
K. Hayashi, Y. Onishi, K. Itoh, H. Miwa, A. Takanishi: Development and evaluation of face robot to express various face shape, Proc. IEEE Int. Conf. Robot. Autom. (2006) pp. 481–486
M. Fujita: On activating human communications with pet-type robot AIBO, Proc. IEEE 92(11), 1804–1813 (2004)
K. Wada, T. Shibata, K. Sakamoto, K. Tanie: Long-term interaction between seal robots and elderly people – robot assisted activity at a health service facility for the aged, Proc. 3rd Int. Symp. Auton. Minirobots Res. Edutain. (2005) pp. 325–330
C.L. Sidner, C. Lee, C.D. Kidd, N. Lesh, C. Rich: Explorations in engagement for humans and robots, Artif. Intell. 166(1-2), 140–164 (2005)
C. Breazeal, A. Brooks, J. Gray, G. Hoffman, C. Kidd, H.J. Lee Lieberman, A. Lockerd, D. Chilongo: Tutelage and collaboration for humanoid robots, Int. J. Humano. Robots 1(2), 315–348 (2004)
M. Mori: Bukimi no tani the uncanny valley, Energy 7(4), 33–35 (1970)
W. Stiehl, J. Lieberman, C. Breazeal, L. Basel, L. Lalla, M. Wolf: Design of a therapeutic robotic companion for relational, affective touch, Proc. 14th IEEE Workshop Robot Human Interact. Commun. (2005) pp. 408–415
H. Kozima: An anthropologist in the childrenʼs world: A field study of childrenʼs everyday interaction with an interactive robot, Proc. Int. Conf. Dev. Learn. (ICDL) (Bloomington 2006)
R. Gockley, R. Simmons, J. Forlizzi: Modeling affect in socially interactive robots, Proc. 15th IEEE Int. Symp. Robot Human Interact. Commun. (2006) pp. 558–563
M.E. Pollack, S. Engberg, J.T. Matthews, S. Thrun, L. Brown, D. Colbry, C. Orosz, B. Peintner, S. Ramakrishnan, J. Dunbar-Jacob, C. McCarthy, M. Montemerlo, J. Pineau, N. Roy: Pearl: A mobile robotic assistant for the elderly, Proc. AAAI Workshop Autom. Eldercare (2002)
Y. Fujita: Personal robot PaPeRo, J. Robot. Mechatron. 14(1), 60–63 (2002)
M.L. Walters, K. Dautenhahn, S.N. Woods, K.L. Koay: Robotic etiquette: Results from user studies involving a fetch and carry task, Proc. 2nd ACM/IEEE Int. Conf. Human-Robot Interact. (Washington 2007) pp. 9–11
R. Gockley, J. Forlizzi, R. Simmons: Natural person-following behavior for social robots, Proc. Human-Robot Interact. (2007) pp. 17–24
A.G. Brooks, C.R. Arkin: Behavioral overlays for non-verbal communication expression on a humanoid robot, Auton. Robots 22(1), 55–74 (2007)
H. Fujisawa, K. Shirai: An algorithm for spoken sentence recognition and its application to the speech input-output system, IEEE Trans. Syst. Man Cybern. SMC-4(5), 475–479 (1974)
T. Kobayashi, Y. Komori, N. Hashimoto, K. Iwata, Y. Fukazawa, J. Yazawa, K. Shirai: Speech conversation system of the musician robot, Proc. ICARʼ85 (1985) pp. 483–488
R. Nisimura, T. Uchida, A. Lee, H. Saruwatari, K. Shikano, Y. Matsumoto: ASKA; receptionist robot with speech dialogue system, Proc. IEEE Int. Conf. Intell. Syst. Robots (2002) pp. 1314–1319
H. Kikuchi, M. Yokoyama, K. Hoashi, Y. Hidaki, T. Kobayashi, K. Shirai: Controlling gaze of humanoid in communication with human, Proc. IROS (1998) pp. 255–260
S. Hashimoto: Humanoid robots in waseda university: Hadaly2 and WABIAN, Auton. Robots 12(1), 25–38 (2002)
C. Breazeal, A. Edsinger, P. Fitzpatrick, B. Scassellati: Active vision systems for sociable robots, IEEE Trans. Syst. Man Cybern. 31(5), 443–453 (2001)
C. Breazeal, C. Kidd, A.L. Thomaz, G. Hoffman, M. Berlin: Effects of nonverbal communication on efficiency and robustness in human-robot teamwork, Proc. Int. Conf. Intell. Robot. Syst. (2005)
Y. Matsusaka, T. Tojo, T. Kobayashi: Conversation robot participating in group conversation, Trans. IEICE E86-D(1), 26–36 (2003)
S. Fujie, K. Fukushima, T. Kobayashi: Back-channel feedback generation using linguistic and nonlinguistic information and its application to spoken dialogue system, Proc. Interspeech (2005) pp. 889–892
M. Imai, T. Ono, H. Ishiguro: Physical relation and expression: Joint attention for human- robot interaction, Proc. RO-MAN2001 (2001) pp. 512–517
S. Fujie, R. Miyake, T. Kobayashi: Spoken dialogue system using recognition of userʼs feedback for rhythmic dialogue, Proc. Speech Prosody (2006)
S. Fujie, Y. Ejiri, K. Nakajima, Y. Matsusaka, T. Kobayashi: A conversation robot using head gesture recognition as paralinguistic information, Proc. IEEE RO-MAN 2004 (2004) pp. 158–164
K. Sakita, K. Ogawara, S. Murakami, K. Kawamura, K. Ikeuchi: Flexible cooperation between human and robot by interpreting human intention from gaze information, Proc. Int. Conf. Intell. Robot Syst. (2004) pp. 846–851
A. Brooks, C. Breazeal: Working with robots and objects: revisiting deictic reference for achieving spatial common ground, Proc. of HRI (Salt Lake City 2006)
P. Cohen, H. Levesque: Teamwork, Nous 25, 487–512 (1991)
H. Clark, D. Wilkes-Gibbs: Referring as a collaborative process, Cognition 22, 1–39 (1986)
F. Tanaka, K. Noda, T. Sawada, M. Fujita: Associated emotion and its expression in an entertainment robot QRIO, Proc. 3rd Int. Conf. Entertain. Comput. (Eindhoven 2004) pp. 499–504
L. Canamero: http://www.felix-growing.org/, February 5 (2007)
C. Breazeal: Function meets style: Insights from emotion theory applied to HRI, IEEE Trans. Syst. Man Cybern. Part C 34(2), 187–194 (2003)
C. Breazeal: Emotion and sociable humanoid robots, Int. J. Human Comput. Interact. 58, 119–155 (2003)
R. Plutchik: Emotions: A general psychoevolutionary theory,. In: Approaches to Emotion, ed. by K. Sherer, P. Elkman (Erlbaum, Hillsdale 1984) pp. 197–219
C. Breazeal, B. Scassellati: Infant-like social interactions between a robot and a human caregiver, Adaptive Behavior 8(1), 47–72 (2000)
C. Breazeal: Early experiments using motivations to regulate human-robot interaction, Proc. 1998 Amer. Assoc. Artif. Intell. (AAAI98) Fall Symposium on Emotional and Intelligent: the Tangled knot of Cognition, Techn. Rep., Vol. FS-98–03 (Orlando 1998) pp. 31–36
C. Breazeal, A. Edsinger, P. Fitzpatrick, B. Scassellati: Active vision systems for sociable robots, IEEE Trans. Syst. Man Cybern. 31(5), 443–453 (2001)
C. Smith, H. Scott: A componential approach to the meaning of facial expressions. In: The Psychology of Facial Expression, ed. by J.J. Russell Fernandez-Dols (Cambridge Univ Press, Cambridge 1997) pp. 229–254
C. Breazeal: Regulation and entrainment for human-robot interaction, Int. J. Exp. Robot. 21(10-11), 883–902 (2003)
C. Breazeal: Emotive qualities in lip synchronized robot speech, Adv. Robot. 17(2), 97–113 (2003)
C. Breazeal, L. Aryananda: Recognizing affective intent in robot directed speech, Auton. Robots 12(1), 85–104 (2002)
A. Fernald: Intonation and communicative intent in motherʼs speech to infants: Is the melody the message?, Child Develop. 60, 1497–1510 (1989)
K. Itoh, H. Miwa, M. Matsumoto, M. Zecca, H. Takanobu, S. Roccella, M.C. Carrozza, P. Dario, A. Takanishi: Behavior model of humanoid robots based on operant conditioning, Proc. 2005 IEEE-RAS Int. Conf. Humano. Robots (Tsukuba 2005) pp. 220–225
H. Miwa, K. Itoh, H. Takanobu, A. Takanishi: Development of mental model for humanoid robots, 15th CISM-IFToMM Symp. Robot Des. Dyn. Contr., Vol. ROM04-10 (Montreal 2004)
H. Miwa, T. Okuchi, K. Itoh, H. Takanobu, A. Takanishi: A new mental model for humanoid robots for human friendly communication: Introduction of learning system, mood vector and second order equations of emotion, Proc. IEEE Int. Conf. Robot. Autom. (2003) pp. 3588–3593
H. Miwa, K. Itoh, K. Ito, H. Takanobu, A. Takanishi: Introduction of the need model for humanoid robots to generate active behavior, Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst. (2003) pp. 1400–1406
P. Ekman, W.V. Friesen: The repertoire of nonverbal behavior, Semiotica 1, 49–98 (1969)
J.A. Russell, M. Bullock: Multidimensional scaling of emotional facial expressions: Similarity from preschool to adults, J. Pers. Social Psychol. 48, 1290–1298 (1985)
C.A. Smith, P.C. Ellsworth: Patterns of cognitive appraisal in emotion, J. Pers. Social Psychol. 48, 813–838 (1985)
B. Scassellati: Foundations for a Theory of Mind for a Humanoid Robot. Ph.D. Thesis (MIT, Department of Computer Science and Electrical Engineering 2000)
A. Leslie: How to acquire a representational theory of mind. In: Metarepresentation: A Multidisciplinary perspective, ed. by D. Sperber (Oxford Univ Press, Oxford 1994) pp. 197–223
S. Baron-Cohen: Precursors to a theory of mind: Understanding attention in others. In: Natural Theories of Mind, ed. by A. Whiten (Blackwell, Oxford 1991) pp. 233–250
B. Scassellati: Mechanisms of shared attention for a humanoid robot, Embodied Cognition and Action: Papers from the 1996 AAAI Fall Symposium (AAAI Menlo Park, 1996)
H. Kozima: Attention-sharing and behavior-sharing in human-robot communication, IEEE Int. Workshop Robot Human Commun. (1998) pp. 9–14
I. Fasel, G.O. Deak, J. Triesch, J. Movellan: Combining embodied models and empirical research for understanding the development of shared attention, Intl. Conf. on Development and Learning (2002) pp. 21–27
G. Butterworth: The ontogeny and phylogeny of joint visual attention. In: Natural Theories of Mind, ed. by A. Whiten (Blackwell, Oxford 1991) pp. 223–232
L.A. Thomaz, M. Berlin, C. Breazeal: An embodied computational model of social referencing, Proc. of 14th IEEE Workshop Robot Human Interact. Commun. (Nashville 2005)
S. Feinman: Social referencing in infancy, Merrill-Palmer Quarterly 28, 445–470 (1982)
F. Strack, L. Martin, S. Stepper: Inhibiting and facilitating conditions of the human smile: A nonobtrusive test of the facial feedback hypothesis, J. Personal. Social Psycholo. 54, 768–777 (1988)
M.P. Niedenthal, W.L. Barsalou, P. Winkielman, S. Krauth-Gruber, F. Ric: Embodiment in attitudes, social perception, and emotion, Personal. Social Psychol. Review 9(3), 184–211 (2005)
C. Breazeal, D. Buchsbaum, J. Gray, D. Gatenby, B. Blumberg: Learning from and about others: Towards using imitation to bootstrap the social understanding of others by robots, Artif. Life 11(1-2), 1–32 (2005), , ed by. L. Rocha and F. Almedia e Costa
M. Davies, T. Stone: Mental Simulation (Blackwell, Oxford 1995)
V. Gallese, A. Goldman: Mirror neurons and the simulation theory of mind-reading, Trends Cognit. Sci. 2(12), 493–501 (1998)
A. Meltzoff, J. Decety: What imitation tells us about social cognition: a rapprochement between developmental psychology and cognitive neuroscience, Philos. Trans. R. Soc. Lond. B Biol. Sci. 358, 491–500 (2003)
L.W. Barsalou, P.M. Niedenthal, A. Barbey, J. Ruppert: Social embodiment. In: The Psychology of Learning and Motivation, Vol. 43, ed. by B. Ross (Academic Press, San Diego 2003) pp. 43–92
M. Johnson, Y. Demiris: Perceptual perspective taking and action recognition, Int. J. Adv. Robot. Syst. 2(4), 301–308 (2005)
J. Gray, C. Breazeal, M. Berlin, A. Brooks, J. Lieberman: Action parsing and goal inference using self as simulator, Proc. 14th IEEE Workshop Robot Human Interact. Commun. (Nashville 2005)
J. Gray, M. Berlin, C. Breazeal: Intention recognition with divergent beliefs for collaborative robots, Proc. AISB 2007: Symp. Mindful Environ. (Newcastle upon Tyne 2007)
J.G. Trafton, N. Cassimatis, M. Bugajska, D. Brock, F. Mintz, A. Schultz: Enabling effective human-robot interaction using perspective-taking in robots, IEEE Trans. Syst. Man Cybern. Part A: Syst. Humans 35(4), 460–470 (2005)
G. Trafton, C.A. Shultz, D. Perzanowsli, W. Adams, M. Bugajska, N. Cassimatis, D. Brock: Children and robots learning to play hide and seek, Proc. 1st Annu. Conf. Human-Robot Interact. (Salt Lake City 2006)
G. Trafton, A. Schultz, M. Bugajska., F. Mintz: Perspective-taking with robots: Experiments and models, Proc. of 14th IEEE Workshop Robot Human Interact. Commun. (Nashville 2005)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2008 Springer-Verlag
About this entry
Cite this entry
Breazeal, C., Takanishi, A., Kobayashi, T. (2008). Social Robots that Interact with People. In: Siciliano, B., Khatib, O. (eds) Springer Handbook of Robotics. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-30301-5_59
Download citation
DOI: https://doi.org/10.1007/978-3-540-30301-5_59
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-23957-4
Online ISBN: 978-3-540-30301-5
eBook Packages: EngineeringEngineering (R0)