Social Robotics

  • Cynthia BreazealEmail author
  • Kerstin Dautenhahn
  • Takayuki Kanda
Part of the Springer Handbooks book series (SHB)


This chapter surveys some of the principal research trends in Social Robotics and its application to human–robot interaction (HRI ). Social (or Sociable) robots are designed to interact with people in a natural, interpersonal manner – often to achieve positive outcomes in diverse applications such as education, health, quality of life, entertainment, communication, and tasks requiring collaborative teamwork. The long-term goal of creating social robots that are competent and capable partners for people is quite a challenging task. They will need to be able to communicate naturally with people using both verbal and nonverbal signals. They will need to engage us not only on a cognitive level, but on an emotional level as well in order to provide effective social and task-related support to people. They will need a wide range of social-cognitive skills and a theory of other minds to understand human behavior, and to be intuitively understood by people. A deep understanding of human intelligence and behavior across multiple dimensions (i. e., cognitive, affective, physical, social, etc.) is necessary in order to design robots that can successfully play a beneficial role in the daily lives of people. This requires a multidisciplinary approach where the design of social robot technologies and methodologies are informed by robotics, artificial intelligence, psychology, neuroscience, human factors, design, anthropology, and more.


Humanoid Robot Social Robot Robot Interaction False Belief Task Ambient Assisted Living 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.



ambient assisted living


activities for daily living


artificial intelligence


automatic spoken-language recognition


Centre for Affective Solutions for Ambient Living Awareness


Carnegie Mellon University


degree of freedom


hierarchical attentive multiple models for execution and recognition


human–robot interaction


Institutional Review Board


Johnson Space Center


liquid-crystal display


National Aeronautics and Space Agency


pleasure arousal dominance


quality of life technology


Robot and Human Interactive Communication


  1. 72.1
    K. Dautenhahn: Getting to know each other – Artificial social intelligence for autonomous robots, Robotics Auton. Syst. 16, 333–356 (1995)CrossRefGoogle Scholar
  2. 72.2
    K. Dautenhahn: I could be you: The phenomenological dimension of social understanding, Cybern. Syst. 28, 417–453 (1997)CrossRefGoogle Scholar
  3. 72.3
    C. Breazeal: Designing Sociable Robots (MIT Press, Cambridge 2002)zbMATHGoogle Scholar
  4. 72.4
    H. Miwa, A. Takanishi, H. Takanobu: Experimental study on robot personality for humanoid head robot, Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2001) pp. 1183–1188Google Scholar
  5. 72.5
    T. Tojo, Y. Matsusaka, T. Ishii, T. Kobayashi: A conversational robot utilizing facial and body expressions, IEEE Int. Conf. Syst. Man Cybern., Vol. 2 (2000) pp. 858–863Google Scholar
  6. 72.6
    J. Cassell, J. Sullivan, S. Prevost, E. Churchill (Eds.): Embodied Conversational Agents (MIT Press, Cambridge 2000)Google Scholar
  7. 72.7
    G. Hoffman, C. Breazeal: Robots that work in collaboration with people, AAAI Symp. Intersect. Cogn. Sci. Robotics, Washington DC (2004)Google Scholar
  8. 72.8
    K. Dautenhahn, A.H. Bond, L. Canamero, B. Edmonds (Eds.): Socially Intelligent Agents: Creating Relationships with Computers and Robots (Kluwer, Boston 2002)Google Scholar
  9. 72.9
    R.W. Picard: Affective Computing (MIT Press, Cambridge 1997)CrossRefGoogle Scholar
  10. 72.10
    T. Fong, I. Nourbakshsh, K. Dautenhahn: A survey of socially interactive robots, Robotics Auton. Syst. 42, 143–166 (2003)zbMATHCrossRefGoogle Scholar
  11. 72.11
    S. Schaal: Is imitation learning the route to humanoid robots?, Trends Cogn. Sci. 3(6), 233–242 (1999)CrossRefGoogle Scholar
  12. 72.12
    J. Fink: Anthropomorphism and human likeness in the design of robots and human-robot interaction, Lect. Notes Comput. Sci. 7621, 199–208 (2012)CrossRefGoogle Scholar
  13. 72.13
    J. Solis, K. Chida, K. Suefuji, A. Takanishi: The development of the anthropomorphic flutist robot at Waseda University, Int. J. Humanoid Robotics 3(2), 1–25 (2006)CrossRefGoogle Scholar
  14. 72.14
    Y. Ogura, H. Aikawa, K. Shimomura, H. Kondo, A. Morishima, H. Lim, A. Takanishi: Development of a new humanoid robot WABIAN-2, Proc. IEEE Int. Conf. Robotics Autom. (ICRA) (2006) pp. 76–81Google Scholar
  15. 72.15
    H. Miwa, K. Itoh, H. Takanobu, A. Takanishi: Mechanical design and motion control of emotion expression Humanoid robot WE-4R, 15th CISM-Symp. Robot Des. Dyn. Control (2004) pp. 255–262Google Scholar
  16. 72.16
    M.F. Jung, J.J. Lee, N. DePalma, S.O. Adalgeirsson, P.J. Hinds, C. Breazeal: Engaging robots: Easing complex human-robot teamwork using backchanneling, Proc. ACM Conf. Comput. Suppor. Coop. Work (CSCW), San Antonio (2013) pp. 1555–1566Google Scholar
  17. 72.17
    J. Goetz, S. Kiesler, A. Powers: Matching robot appearance and behavior to tasks to improve human-robot cooperation, Proc. 12th IEEE Int. Workshop Robot Hum. Interact. Commun. (ROMAN) (2003) pp. 55–60Google Scholar
  18. 72.18
    F. Iida, M. Tabata, F. Hara: Generating personality character in a face robot through interaction with human, Proc. 7th IEEE Int. Workshop Robot Hum. Commun. (ROMA) (1998) pp. 481–486Google Scholar
  19. 72.19
    M. Shimada, T. Minato, S. Itakura, H. Ishiguro: Evaluation of android using unconscious recognition, Proc. 6th IEEE-RAS Int. Conf. Humanoid Robots (2006) pp. 157–162Google Scholar
  20. 72.20
    K. Berns, J.J. Hirth: Control of facial expressions of the humanoid robot head ROMAN, Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), Bejing (2006) pp. 3119–3124Google Scholar
  21. 72.21
    K. Dautenhahn, C.L. Nehaniv, M.L. Walters, B. Robins, H. Kose-Bagci, N.A. Mirza, M. Blow: KASPAR - A minimally expressive humanoid robot for human-robot interaction research, Appl. Bionics Biomech. 6(3), 369–397 (2009)CrossRefGoogle Scholar
  22. 72.22
    F. Tanaka, J.R. Movellan, B. Fortenberry, K. Aisaka: Daily HRI evaluation at a classroom environment: Reports from dance interaction experiments, Proc. 1st ACM Conf. Hum.-Robot Interact. (HRI), Salt Lake City (2006) pp. 3–9Google Scholar
  23. 72.23
    H. Lim, S. Hyon, S.A. Setiawan, A. Takanishi: Quasi-human biped walking, Int. J. Inform. Educ. Res. Robotics Artif. Intell. 24(2), 257–268 (2006)Google Scholar
  24. 72.24
    H. Lim, A. Ishii, A. Takanishi: Emotion-based biped walking, Int. J. Inform. Educ. Res. Robotics Artif. Intell. 22(5), 577–586 (2004)zbMATHGoogle Scholar
  25. 72.25
    J. Solis, S. Isoda, K. Chida, A. Takanishi, K. Wakamatsu: Anthropomorphic flutist robot for teaching flute playing to beginner students, Proc. IEEE Int. Conf. Robotics Autom. (ICRA) (2004) pp. 146–151Google Scholar
  26. 72.26
    H. Miwa, K. Itoh, M. Matsumoto, M. Zecca, H. Takanobu, S. Roccella, M.C. Carrozza, P. Dario, A. Takanishi: Effective emotional expressions with emotion expression humanoid robot WE-4RII, Proc. 2004 IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2004) pp. 2203–2208Google Scholar
  27. 72.27
    S. Roccella, M.C. Carrozza, G. Cappiello, P. Dario, J. Cabibihan, M. Zecca, H. Miwa, K. Itoh, M. Matsumoto, A. Takanishi: Design, fabrication and preliminary results of a novel anthropomorphic hand for humanoid robotics: RCH-1, Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), Sendai (2004) pp. 266–271Google Scholar
  28. 72.28
    H. Miwa, K. Itoh, H. Takanobu, A. Takanishi: Design and control of 9-DOFs emotion expression humanoid arm, Proc. IEEE Int. Conf. Robotics Autom. (ICRA), New Orleans (2004) pp. 128–133Google Scholar
  29. 72.29
    K. Hayashi, Y. Onishi, K. Itoh, H. Miwa, A. Takanishi: Development and evaluation of face robot to express various face shape, Proc. IEEE Int. Conf. Robotics Autom. (ICRA) (2006) pp. 481–486Google Scholar
  30. 72.30
    M. Fujita: On activating human communications with pet-type robot AIBO, Proceedings IEEE 92(11), 1804–1813 (2004)CrossRefGoogle Scholar
  31. 72.31
    K. Wada, T. Shibata, K. Sakamoto, K. Tanie: Long-term interaction between seal robots and elderly people — Robot assisted activity at a health service facility for the aged, Proc. 3rd Int. Symp. Auton. Minirobots Res. Edutainment (2005) pp. 325–330Google Scholar
  32. 72.32
    C.L. Sidner, C. Lee, C.D. Kidd, N. Lesh, C. Rich: Explorations in engagement for humans and robots, Artif. Intell. 166(1/2), 140–164 (2005)CrossRefGoogle Scholar
  33. 72.33
    C. Breazeal, A. Brooks, J. Gray, G. Hoffman, C. Kidd, H. Lee, J. Lieberman, A. Lockerd, D. Chilongo: Tutelage and collaboration for humanoid robots, Int. J. Humanoid Robotics 1(2), 315–348 (2004)CrossRefGoogle Scholar
  34. 72.34
    H. Kozima: An anthropologist in the children’s world: A field study of children’s everyday interaction with an interactive robot, Proc. Int. Conf. Dev. Learn. (ICDL), Bloomington (2006)Google Scholar
  35. 72.35
    M.E. Pollack, S. Engberg, J.T. Matthews, S. Thrun, L. Brown, D. Colbry, C. Orosz, B. Peintner, S. Ramakrishnan, J. Dunbar-Jacob, C. McCarthy, M. Montemerlo, J. Pineau, N. Roy: Pearl: A mobile robotic assistant for the elderly, Proc. AAAI Workshop Autom. Eldercare (2002)Google Scholar
  36. 72.36
    R. Gockley, R. Simmons, J. Forlizzi: Modeling affect in socially interactive robots, Proc. 15th IEEE Int. Symp. Robot Hum. Interact. Commun. (ROMAN) (2006) pp. 558–563Google Scholar
  37. 72.37
    M. Mori: Bukimi no tani the uncanny valley, Energy 7(4), 33–35 (1970)Google Scholar
  38. 72.38
    K. Wada, T. Shibata: Living with seal robots in a care house – Evaluations of social and physiological influences, IEEE/RSJ Int. Conf. Intell. Robots Syst. (2007) pp. 4940–4945Google Scholar
  39. 72.39
    W. Stiehl, J. Lieberman, C. Breazeal, L. Basel, L. Lalla, M. Wolf: Design of a therapeutic robotic companion for relational, affective touch, Proc. 14th IEEE Workshop Robot Hum. Interact. Commun. (ROMAN) (2005) pp. 408–415Google Scholar
  40. 72.40
    NEC Corporation Japan:
  41. 72.41
    M.L. Walters, K. Dautenhahn, S.N. Woods, K.L. Koay: Robotic etiquette: Results from user studies involving a fetch and carry task, Proc. 2nd ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI), Washington (2007)Google Scholar
  42. 72.42
    R. Gockley, J. Forlizzi, R. Simmons: Natural person-following behavior for social robots, Proc. 2nd ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2007) pp. 17–24Google Scholar
  43. 72.43
    A.G. Brooks, R.C. Arkin: Behavioral overlays for non-verbal communication expression on a humanoid robot, Auton. Robots 22(1), 55–74 (2007)CrossRefGoogle Scholar
  44. 72.44
    C. Darwin, P. Ekman: The Expression of the Emotions in Man and Animals, 3rd edn. (Oxford Univ. Press, Oxford 1998)Google Scholar
  45. 72.45
    C. Breazeal: Emotion and sociable humanoid robots, Int. J. Hum. Comput. Interact. 58, 119–155 (2003)Google Scholar
  46. 72.46
    F. Tanaka, K. Noda, T. Sawada, M. Fujita: Associated emotion and its expression in an entertainment robot QRIO, Proc. 3rd Int. Conf. Entertain. Comput., Eindhoven (2004) pp. 499–504Google Scholar
  47. 72.47
    S. Fujie, Y. Ejiri, K. Nakajima, Y. Matsusaka, T. Kobayashi: A conversation robot using head gesture recognition as paralinguistic information, Proc. IEEE Int. Symp. Robot Hum. Interact. Commun. (ROMAN) (2004) pp. 158–164Google Scholar
  48. 72.48
    L. Hall, S. Woods, R. Aylett, L. Newall, A. Paiva: Achieving empathic engagement through affective interaction with synthetic characters, Lect. Notes Comput. Sci. 3784, 731–738 (2005)CrossRefGoogle Scholar
  49. 72.49
    R.C. Arkin: Moving up the food chain: Motivation and emotion in behavior-based robots. In: Who Needs Emotions: The Brain Meets the Robot, ed. by J. Fellous, M. Arbib (Oxford Univ. Press, Oxford 2005)Google Scholar
  50. 72.50
    Feelix growing, a European project coordinated by University of Hertfordshire:
  51. 72.51
    C. Breazeal: Function meets style: Insights from emotion theory applied to HRI, IEEE Trans. Syst. Man Cybern. C 34(2), 187–194 (2003)CrossRefGoogle Scholar
  52. 72.52
    A. Sloman: Beyond shallow models of emotion, Cogn. Process. Int. Q. Cogn. Sci. 2(1), 177–198 (2001)Google Scholar
  53. 72.53
    S. Chernova, N. DePalma, C. Breazeal: Crowdsourcing real world human-robot dialog and teamwork through online multiplayer games, AAAI Magazine 32(4), 100–111 (2011)Google Scholar
  54. 72.54
    M. Scheutz, P. Schermerhorn, J. Kramer: The utility of affect expression in natural language interactions in joint human-robot tasks, Proc. 1st ACM Conf. Hum.-Robot Interact. (HRI) (ACM, New York, USA 2006) pp. 226–233Google Scholar
  55. 72.55
    J.C. Lester, S.G. Towns, C.B. Callaway, J.L. Voerman, P.J. Fitzgerald: Deictic and emotive communication in animated pedagogical agents. In: Embodied Conversational Agents, ed. by J. Casell, S. Prevost, J. Sullivan, E. Churchill (MIT Press, Cambridge 2000)Google Scholar
  56. 72.56
    M.E. Hoque: My automated conversation helper (MACH): Helping people improve social skills, Proc. 14th ACM Int. Conf. Multimodal Interact. (ICMI), Santa Monica (2012)Google Scholar
  57. 72.57
    R. Lazarus: Emotion and Adaptation (Oxford Univ. Press, New York 1991)Google Scholar
  58. 72.58
    A. Ortony, G. Clore, A. Collins: The Cognitive Structure of Emotions (Cambridge Univ. Press, New York 1988)CrossRefGoogle Scholar
  59. 72.59
    A. Marsella, J. Gratch: EMA: A process model of appraisal dynamics, J. Cogn. Syst. Res. 10, 70–90 (2009)CrossRefGoogle Scholar
  60. 72.60
    C. Elliott: The Affective Reasoner: A Process Model of Emotions in a Multi-Agent System, Ph.D. Thesis (Northwestern Univ., Northwestern 1992)Google Scholar
  61. 72.61
    J.A. Russell: Core affect and the psychological construation of emotion, Psychol. Rev. 110, 145–172 (2003)CrossRefGoogle Scholar
  62. 72.62
    A. Mehrabian, J.A. Russell: An Approach to Environmental Psychology (MIT Press, Cambridge 1974)Google Scholar
  63. 72.63
    C. Smith, H. Scott: A componential approach to the meaning of facial expressions. In: The Psychology of Facial Expression, ed. by J. Russell, J. Fernandez-Dols (Cambridge Univ. Press, Cambride 1997) pp. 229–254CrossRefGoogle Scholar
  64. 72.64
    C. Breazeal, B. Scassellati: Infant-like social interactions between a robot and a human caregiver, Adapt. Behav. 8(1), 47–72 (2000)CrossRefGoogle Scholar
  65. 72.65
    A. Damasio: Decartes’ Error: Emotion, Reason and the Human Brain (Putnam, New York 1994)Google Scholar
  66. 72.66
    C. Breazeal: Emotive qualities in lip synchronized robot speech, Adv. Robotics 17(2), 97–113 (2003)CrossRefGoogle Scholar
  67. 72.67
    C. Breazeal: Early experiments using motivations to regulate human-robot interaction, AAAI Fall Symp. Emot. Intell., Orlando (1998) pp. 31–36Google Scholar
  68. 72.68
    C. Breazeal, L. Aryananda: Recognizing affective intent in robot directed speech, Auton. Robots 12(1), 85–104 (2002)zbMATHCrossRefGoogle Scholar
  69. 72.69
    J. Panskepp: Affective Neuroscience: The Foundations of Human and Animal Emotions (Oxford Univ. Press, New York 1998)Google Scholar
  70. 72.70
    R. Plutchik: Emotions: A general psychoevolutionary theory. In: Approaches to Emotion, ed. by K. Sherer, P. Elkman (Lawrence Erlbaum Associates, Hillsdale 1984) pp. 197–219Google Scholar
  71. 72.71
    L. Vygotsky: Mind in Society: The Development of Higher Psychological Processes (Harvard Univ. Press, Cambridge 1978)Google Scholar
  72. 72.72
    F. Strack, L. Martin, S. Stepper: Inhibiting and facilitating conditions of the human smile: A nonobtrusive test of the facial feedback hypothesis, J. Person. Soc. Psychol. 54, 768–777 (1988)CrossRefGoogle Scholar
  73. 72.73
    P.M. Niedenthal, L.W. Barsalou, P. Winkielman, S. Krauth-Gruber, F. Ric: Embodiment in attitudes, social perception, and emotion, Personal. Soc. Psychol. Rev. 9(3), 184–211 (2005)CrossRefGoogle Scholar
  74. 72.74
    C. Breazeal, D. Buchsbaum, J. Gray, D. Gatenby, B. Blumberg: Learning from and about others: Towards using imitation to bootstrap the social understanding of others by robots, Artif. Life 11(1–2), 31–62 (2005)CrossRefGoogle Scholar
  75. 72.75
    A. Fernald: Intonation and communicative intent in mother’s speech to infants: Is the melody the message?, Child Dev. 60, 1497–1510 (1989)CrossRefGoogle Scholar
  76. 72.76
    B. Scassellati: Theory of mind for a humanoid robot, Auton. Robots 12(1), 13–24 (2002)zbMATHCrossRefGoogle Scholar
  77. 72.77
    A. Leslie: How to acquire a representational theory of mind. In: Metarepresentation: A Multidisciplinary Perspective, ed. by D. Sperber (Oxford Univ. Press, Oxford 1994) pp. 197–223Google Scholar
  78. 72.78
    S. Baron-Cohen: Precursors to a theory of mind: Understanding attention in others. In: Natural Theories of Mind, ed. by A. Whiten (Blackwell, Oxford 1991) pp. 233–250Google Scholar
  79. 72.79
    B. Scassellati: Mechanisms of shared attention for a humanoid robot, AAAI Fall Symp. Embodied Cogn. Action (1996)Google Scholar
  80. 72.80
    G.O. Fasel, J. Deak, J. Triesch, J. Movellan: Combining embodied models and empirical research for understanding the development of shared attention, Proc. 2nd IEEE Int. Conf. Dev. Learn. (ICDL) (2002) pp. 21–27Google Scholar
  81. 72.81
    Y. Nagai, M. Asada, K. Hosoda: Learning for joint attention helped by functional development, Adv. Robotics 20(10), 1165–1181 (2006)CrossRefGoogle Scholar
  82. 72.82
    G. Butterworth: The ontogeny and phylogeny of joint visual attention. In: Natural Theories of Mind, ed. by A. Whiten (Blackwell, Oxford 1991) pp. 223–232Google Scholar
  83. 72.83
    A.L. Thomaz, M. Berlin, C. Breazeal: An embodied computational model of social referencing, Proc. 14th IEEE Workshop Robot Hum. Interact. Commun. (ROMAN), Nashville (2005)Google Scholar
  84. 72.84
    S. Feinman: Social referencing in infancy, Merrill-Palmer Q. 28, 445–470 (1982)Google Scholar
  85. 72.85
    M. Davies, T. Stone: Mental Simulation (Blackwell, Oxford 1995)Google Scholar
  86. 72.86
    V. Gallese, A. Goldman: Mirror neurons and the simulation theory of mind-reading, Trends Cogn. Sci. 2(12), 493–501 (1998)CrossRefGoogle Scholar
  87. 72.87
    A. Meltzoff, J. Decety: What imitation tells us about social cognition: A rapprochement between developmental psychology and cognitive neuroscience, Philos. Trans. R. Soc. Lond. B Biol, Sci. 358, 491–500 (2003)Google Scholar
  88. 72.88
    L.W. Barsalou, P.M. Niedenthal, A. Barbey, J. Ruppert: Social embodiment. In: The Psychology of Learning and Motivation, Vol. 43, ed. by B. Ross (Academic, Amsterdam 2003) pp. 43–92Google Scholar
  89. 72.89
    M. Johnson, Y. Demiris: Perceptual perspective taking and action recognition, Int. J. Adv. Robotics Syst. 2(4), 301–308 (2005)Google Scholar
  90. 72.90
    J. Gray, C. Breazeal, M. Berlin, A. Brooks, J. Lieberman: Action parsing and goal inference using self as simulator, Proc. 14th IEEE Workshop Robot Hum. Interact. Commun. (ROMAN), Nashville (2005)Google Scholar
  91. 72.91
    J. Gray, M. Berlin, C. Breazeal: Intention recognition with divergent beliefs for collaborative robots, Proc. AISB Symp. Mindful Environ., Newcastle Upon Tyne (2007)Google Scholar
  92. 72.92
    G. Trafton, A. Schultz, M. Bugajska, F. Mintz: Perspective-taking with Robots: Experiments and models, Proc. 14th IEEE Workshop Robot Hum. Interact. Commun. (ROMAN), Nashville (2005)Google Scholar
  93. 72.93
    G. Trafton, A.C. Shultz, D. Perzanowsli, W. Adams, M. Bugajska, N. Cassimatis, D. Brock: Children and robots learning to play hide and seek, Proc. 1st Ann. Conf. Hum.-Robot Interact. (HRI), Salt Lake City (2006)Google Scholar
  94. 72.94
    J.G. Trafton, N. Cassimatis, M. Bugajska, D. Brock, F. Mintz, A. Schultz: Enabling effective human-robot interaction using perspective-taking in robots, IEEE Trans. Syst. Man Cybern. A Syst. Hum. 35(4), 460–470 (2005)CrossRefGoogle Scholar
  95. 72.95
    J. Gray, C. Breazeal: Manipulating mental states through physical action, Int. Conf. Soc. Robotics, Chendu (2012)Google Scholar
  96. 72.96
    C. Nass, S. Brave: Wired for Speech: How Voice Activates and Advances the Human-Computer Relationship (MIT Press, Cambridge, MA 2005)Google Scholar
  97. 72.97
    C. Nass, Y. Moon: Machines and mindlessness: Social responses to computers, J. Soc. Issues 56, 81–103 (2000)CrossRefGoogle Scholar
  98. 72.98
    C. Nass, J.S. Steuer, E. Tauber, H. Reeder: Anthropomorphism, agency, and ethopoeia: Computers as social actors, ACM Conf. Companion Hum. Factors Comput. Syst., Amsterdam (1993) pp. 111–112Google Scholar
  99. 72.99
    N. Epley, A. Waytz, J.T. Cacioppo: On seeing human: A three-factor theory of anthropomorphism, Psychol. Rev. 114, 864–886 (2007)CrossRefGoogle Scholar
  100. 72.100
    J. Goetz, S. Kiesler: Cooperation with a robotic assistant, ACM Ext. Abstr. Hum. Factors Comput. Syst. (2002) pp. 578–579Google Scholar
  101. 72.101
    L. Axelrod, K. Hone: E-motional advantage: Performance and satisfaction gains with affective computing, ACM Ext. Abstr. Hum. Factors Comput. Sci. (2005) pp. 1192–1195Google Scholar
  102. 72.102
    S.O. Adalgeirsson, C. Breazeal: MeBot: A robotic platform for socially embodied presence, Proc. 5th ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2010) pp. 15–22Google Scholar
  103. 72.103
    S. Lee, I.Y. Lau, S. Kiesler, C.-Y. Chiu: Human mental models of humanoid robots, Proc. IEEE Int. Conf. Robotics Autom. (ICRA) (2005) pp. 2767–2772Google Scholar
  104. 72.104
    L.D. Riek, T.-C. Rabinowitch, B. Chakrabarti, P. Robinson: How anthropomorphism affects empathy toward robots, Proc. 4th ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2009) pp. 245–246CrossRefGoogle Scholar
  105. 72.105
    S. Krach, F. Hegel, B. Wrede, G. Sagerer, F. Binkofski, T. Kircher: Can machines think? Interaction and perspective taking with robots investigated via fMRI, PLoS ONE 3(7), e2597 (2008)CrossRefGoogle Scholar
  106. 72.106
    E. Short, J. Hart, M. Vu, B. Scassellati: No fair!! An interaction with a cheating robot, Proc. 5th ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2010) pp. 219–226Google Scholar
  107. 72.107
    S.R. Fussell, S. Kiesler, L.D. Setlock, V. Yew: How people anthropomorphize robots, Proc. 3rd ACM/IEEE Int. Conf. Human Robot Interact. (HRI) (2008) pp. 145–152Google Scholar
  108. 72.108
    P.H. Kahn Jr., N.G. Freier, B. Friedman, R.L. Severson, E. Feldman: Social and moral relationships with robotic others?, Proc. 13th Int. Workshop Robot Hum. Interact. Commun. (ROMAN) (2004) pp. 545–550Google Scholar
  109. 72.109
    C. Bartneck, J. Forlizzi: A design-centred framework for social human-robot interaction, 13th IEEE Int. Workshop Robot Hum. Interact. Commun. (ROMAN) (2004) pp. 591–594Google Scholar
  110. 72.110
    J. Shah, C. Breazeal: Improved human-robot team performance using chaski, A human-inspired plan execution system, Proc. ACM/IEEE Int. Conf. Hum. Robot Interact. (HRI) (2011)Google Scholar
  111. 72.111
    Rethink Robotics:
  112. 72.112
    C. Kidd, C. Breazeal: Effect of a robot on user perceptions, Proc. 2004 IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), Sendai, Vol. 4 (2004) pp. 3559–3564Google Scholar
  113. 72.113
    W. Bainbridge, J. Hart, E. Kim, B. Scassellati: The effect of presence on human-robot interaction, IEEE Int. Symp. Robot Hum. Interact. Commun., Munich (2008)Google Scholar
  114. 72.114
    J. Wainer, D. Feil-Seifer, D. Shell, M. Matarić: The role of physical embodiment in human-robot interaction, Proc. Int. Workshop Robot Hum. Interact. Commun. (ROMAN) (2006) pp. 6–8Google Scholar
  115. 72.115
    J. Wainer, D. Feil-Seifer, D. Shell, M. Matarić: Embodiment and human-robot interaction: A task-based perspective, Proc. 16th IEEE Int. Workshop Robot Hum. Interact. Commun. (ROMAN), Jeju Island (2007)Google Scholar
  116. 72.116
    C. Kidd, C. Breazeal: Robots at home: Understanding long-term human-robot interaction, Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), Nice (2008)Google Scholar
  117. 72.117
    F.R. Volkmar, C. Lord, A. Bailey, R.T. Schultz, A. Klin: Autism and pervasive developmental disorders, J. Child Psychol. Psych. 45(1), 1–36 (2004)CrossRefGoogle Scholar
  118. 72.118
    N. Freed: This is the Fluffy Robot That Only Speaks French: Language Use Between Preschoolers, Their Families, and a Social Robot While Sharing Virtual Toys. Masters Sci. Thesis (MIT, Cambridge 2012)Google Scholar
  119. 72.119
    S. Turkle: Along Together (Basic Books, New York 2012)Google Scholar
  120. 72.120
    N. Ambady, M. Weisbuch: Nonverbal behavior. In: Handbook of Social Psychology, 5th edn., ed. by D.T. Gilbert, S.T. Fiske, G. Lindzey (Wiley, Hoboken 2010)Google Scholar
  121. 72.121
    J.A. Hall, E.J. Coats, L. Smith-Lebeau: Nonverbal behavior and the vertical dimension of social relations: A meta-analysis, Psychol. Bull. 131, 898–924 (2005)CrossRefGoogle Scholar
  122. 72.122
    D. DeSteno, C. Breazeal, R. Frank, D. Pizarro, J. Baumann, L. Dickens, J.J. Lee: Detecting the trustworthiness of novel partners in economic exchange, Psychol. Sci. 23(12), 1549–1556 (2012)CrossRefGoogle Scholar
  123. 72.123
    C.P.H. Langford, J. Bowsher, J.P. Maloney, P.P. Lillis: Social support: A conceptual analysis, J. Adv. Nurs. 25, 95–100 (1997)CrossRefGoogle Scholar
  124. 72.124
    T.A. Wills: Supportive functions of interpersonal relationships. In: Social Support and Health, ed. by S. Cohen, L. Syme (Academic, Orlando 1985) pp. 61–82Google Scholar
  125. 72.125
    T.A. Wills: Social support and interpersonal relationships. In: Prosocial Behavior, ed. by M.S. Clark (Sage, Newbury Park 1991) pp. 265–289Google Scholar
  126. 72.126
    F.J. Bernieri, J.M. Davis, R. Rosenthal, C.R. Knee: Inter-actional synchrony and rapport: Measuring synchrony in displays devoid of sound and facial affect, Personal. Soc. Psychol. Bull. 20, 303–311 (1994)CrossRefGoogle Scholar
  127. 72.127
    F.J. Bernieri: Coordinated movement and rapport in teacher-student interactions, J. Nonverbal Behav. 12, 120–138 (1988)CrossRefGoogle Scholar
  128. 72.128
    N. Sonalkar, M. Jung, A. Mabogunje: Emotion in engineering design teams. In: Emotional Engineering: Service Development, ed. by S. Fukuda (Springer, London 2010)Google Scholar
  129. 72.129
    J.E. Grahe, F.J. Bernieri: The importance of nonverbal cues in judging rapport, J. Nonverbal Behav. 23, 253–269 (1999)CrossRefGoogle Scholar
  130. 72.130
    L.K. Miles, L.K. Nind, C.N. Macrae: The rhythm of rapport: Interpersonal synchrony and social perception, J. Exp. Soc. Psychol. 45(3), 585–589 (2009)CrossRefGoogle Scholar
  131. 72.131
    Y. Nagai: Learning to comprehend deictic gestures in robots and human infants, IEEE Int. Workshop Robot Hum. Interact. Commun. (ROMAN) (2005) pp. 217–222Google Scholar
  132. 72.132
    Z. Kasap, N. Magnenat-Thalmann: Building long-term relationships with virtual and robotic characters: The role of remembering, Vis. Comput. 28, 87–97 (2012)CrossRefGoogle Scholar
  133. 72.133
    F. Tanaka, S. Matsuzoe: Children teach a care-receiving robot to promote their learning: Field experiments in a classroom for vocabulary learning, J. Hum.-Robot Interact. 1(1), 78–95 (2012)CrossRefGoogle Scholar
  134. 72.134
    C. Pastor, G. Gaminde, A. Renteria: COMPANIONABLE: Integrated cognitive assistive and domotic companion robotic systems for ability and security, Int. Symp. Robotics, Barcelona (2009)Google Scholar
  135. 72.135
    H. Fujisawa, K. Shirai: An algorithm for spoken sentence recognition and its application to the speech input-output system, Proc. IEEE Trans. Syst. Man Cybern. 4(5), 475–479 (1974)zbMATHGoogle Scholar
  136. 72.136
    T. Kobayashi, Y. Komori, N. Hashimoto, K. Iwata, Y. Fukazawa, J. Yazawa, K. Shirai: Speech conversation system of the musician robot, Proc. Int. Conf. Adv. Robotics (ICAR) (1985) pp. 483–488Google Scholar
  137. 72.137
    H. Kikuchi, M. Yokoyama, K. Hoashi, Y. Hidaki, T. Kobayashi, K. Shirai: Controlling gaze of humanoid in communication with human, Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (1998) pp. 255–260Google Scholar
  138. 72.138
    S. Hashimoto: Humanoid robots in Waseda University: Hadaly2 and WABIAN, Auton. Robots 12(1), 25–38 (2002)zbMATHCrossRefGoogle Scholar
  139. 72.139
    C. Rich, B. Ponsler, A. Holroyd, C.L. Sidner: Recognizing engagement in human-robot interaction, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2010) pp. 375–382Google Scholar
  140. 72.140
    C. Breazeal, A. Edsinger, P. Fitzpatrick, B. Scassellati: Active vision systems for sociable robots, IEEE Trans. Syst. Man Cybern. 31(5), 443–453 (2001)CrossRefGoogle Scholar
  141. 72.141
    C. Breazeal, C.D. Kidd, A.L. Thomaz, G. Hoffman, M. Berlin: Effects of nonverbal communication on efficiency and robustness in human-robot teamwork, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2005) pp. 383–388Google Scholar
  142. 72.142
    C. Breazeal, C. Kidd, A.L. Thomaz, G. Hoffman, M. Berlin: Effects of nonverbal communication on efficiency and robustness in human-robot teamwork, Proc. IEEE Int. Conf. Intell. Robots Syst. (IROS) (2005)Google Scholar
  143. 72.143
    B. Mutlu, T. Shiwa, T. Kanda, H. Ishiguro, N. Hagita: Footing in human-robot conversations: How robots might shape participant roles using gaze cues, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2009) pp. 61–68Google Scholar
  144. 72.144
    B. Mutlu, F. Yamaoka, T. Kanda, H. Ishiguro, N. Hagita: Nonverbal leakage in robots: Communication of intentions through seemingly unintentional behavior, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2009) pp. 69–76Google Scholar
  145. 72.145
    N. Kirchner, A. Alempijevic: A robot centric perspective on HRI, J. Hum.-Robot Interact. 1(2), 135–157 (2012)Google Scholar
  146. 72.146
    A. Yamazaki, K. Yamazaki, T. Ohyama, Y. Kobayashi, Y. Kuno: A techno-sociological solution for designing a Museum guide robot: Regarding choosing an appropriate visitor, Proc. 17th Annu. ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2012) pp. 309–316Google Scholar
  147. 72.147
    S. Fujie, K. Fukushima, T. Kobayashi: Back-channel feedback generation using linguistic and nonlinguistic information and its application to spoken dialogue system, Proc. Interspeech (2005) pp. 889–892Google Scholar
  148. 72.148
    M. Imai, T. Ono, H. Ishiguro: Physical relation and expression: Joint attention for human-robot interaction, Proc. ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2001) pp. 512–517Google Scholar
  149. 72.149
    T. Kanda, M. Kamasima, M. Imai, T. Ono, D. Sakamoto, H. Ishiguro, Y. Anzai: A humanoid robot that pretends to listen to route guidance from a human, Auton. Robots 22, 87–100 (2007)CrossRefGoogle Scholar
  150. 72.150
    H. Ogawa, T. Watanabe: Interrobot: A speech driven embodied interaction robot, IEEE Int. Workshop Robot Hum. Interact. Commun. (ROMAN) (2000) pp. 322–327Google Scholar
  151. 72.151
    D. Sakamoto, T. Kanda, T. Ono, H. Ishiguro, N. Hagita: Android as a telecommunication medium with a human-like presence, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2007) pp. 193–200Google Scholar
  152. 72.152
    H. Kuzuoka, S. Oyama, K. Yamazaki, K. Suzuk, M. Mitsuishi: GestureMan: A mobile robot that embodies a remote instructor’s actions, Proc. ACM Conf. Comput.-Suppor. Coop. Work (CSCW) (2000)Google Scholar
  153. 72.153
    A. Brooks, C. Breazeal: Working with robots and objects: Revisiting deictic reference for achieving spatial common ground, Proc. ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2006)Google Scholar
  154. 72.154
    B. Scassellati: Investigating Models of Social Development Using a Humanoid Robot. Biorobotics (MIT Press, Cambridge 2000)Google Scholar
  155. 72.155
    H. Kozima, E. Vatikiotis-Bateson: Communicative criteria for processing time/space-varying information, IEEE Int. Workshop Robot Hum. Commun. (ROMAN) (2001)Google Scholar
  156. 72.156
    O. Sugiyama, T. Kanda, M. Imai, H. Ishiguro, N. Hagita: Humanlike conversation with gestures and verbal cues based on a three-layer attention-drawing model, Connect. Sci. 18(4), 379–402 (2006)CrossRefGoogle Scholar
  157. 72.157
    V. Ng-Thow-Hing, P. Luo, S. Okita: Synchronized gesture and speech production for humanoid robots, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2010) pp. 4617–4624Google Scholar
  158. 72.158
    Y. Hato, S. Satake, T. Kanda, M. Imai, N. Hagita: Pointing to space: Modeling of deictic interaction referring to regions, Proc. ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2010) pp. 301–308Google Scholar
  159. 72.159
    Y. Okuno, T. Kanda, M. Imai, H. Ishiguro, N. Hagita: Providing route directions: Design of robot’s utterance, gesture, and timing, Proc. ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2009) pp. 53–60CrossRefGoogle Scholar
  160. 72.160
    O. Sugiyama, T. Kanda, M. Imai, H. Ishiguro, N. Hagita: Three-layer model for generation and recognition of attention-drawing behavior, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2006) pp. 5843–5850Google Scholar
  161. 72.161
    H. Sacks, E.A. Schegloff, G. Jefferson: A simplest systematics for the organization of turn-taking for conversation, Language 50, 696–735 (1974)CrossRefGoogle Scholar
  162. 72.162
    H.H. Clark: Using Language (Cambridge Univ. Press, Cambridge 1996)CrossRefGoogle Scholar
  163. 72.163
    M.F. McTear: Spoken dialogue technology: Enabling the conversational user interface, ACM Comput. Surv. 34, 90–169 (2002)CrossRefGoogle Scholar
  164. 72.164
    M. Nakano, Y. Hasegawa, K. Nakadai, T. Nakamura, J. Takeuchi, T. Torii, H. Tsujino, N. Kanda, H.G. Okuno: A two-layer model for behavior and dialogue planning in conversational service robots, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2005) pp. 3329–3335Google Scholar
  165. 72.165
    M. Scheutz, P. Schermerhorn, J. Kramer, D. Anderson: First steps toward natural human-like Hri, Auton. Robots 22, 411–423 (2006)CrossRefGoogle Scholar
  166. 72.166
    C. Shi, T. Kanda, M. Shimada, F. Yamaoka, H. Ishiguro, N. Hagita: Easy development of communicative behaviors in social robots, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2010) pp. 5302–5309Google Scholar
  167. 72.167
    K. Sakita, K. Ogawara, S. Murakami, K. Kawamura, K. Ikeuchi: Flexible cooperation between human and robot by interpreting human intention from gaze information, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2004) pp. 846–851Google Scholar
  168. 72.168
    Y. Matsusaka, T. Tojo, T. Kobayashi: Conversation robot participating in group conversation, IEICI Trans. Inform. Syst. 86(1), 26–36 (2003)Google Scholar
  169. 72.169
    R. Nisimura, T. Uchida, A. Lee, H. Saruwatari, K. Shikano, Y. Matsumoto: ASKA: Receptionist robot with speech dialogue system, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2002) pp. 1314–1319CrossRefGoogle Scholar
  170. 72.170
    Y. Matsusaka, S. Fujie, T. Kobayashi: Modeling of conversational strategy for the robot participating in the group conversation, Eur. Conf. Speech Commun. Technol. (EUROSPEECH) (2001)Google Scholar
  171. 72.171
    M. Hoque, D. Das, T. Onuki, Y. Kobayashi, Y. Kuno: An integrated approach of attention control of target human by nonverbal behaviors of robots in different viewing situations, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2012) pp. 1399–1406Google Scholar
  172. 72.172
    A. Raux, M. Eskenazi: A finite-state turn-taking model for spoken dialog systems, Proc. Hum. Lang. Technol. Annu. Conf. North Am. Chapt. Assoc. Comput. Linguist. (2009) pp. 629–637Google Scholar
  173. 72.173
    C. Chao, A.L. Thomaz: Timing in multimodal turn-taking interactions: Control and analysis using timed petri nets, J. Hum.-Robot Interact. 1, 4–25 (2012)CrossRefGoogle Scholar
  174. 72.174
    M. Yamamoto, T. Watanabe: Time lag effects of utterance to communicative actions on Cg character-human greeting interaction, IEEE Int. Symp. Robot Hum. Interact. Commun. (ROMAN) (2006) pp. 629–634Google Scholar
  175. 72.175
    D. McNeill: Psycholinguistics: A New Approach (HarperRow, New York 1987)zbMATHGoogle Scholar
  176. 72.176
    T. Shiwa, T. Kanda, M. Imai, H. Ishiguro, N. Hagita: How quickly should a communication robot respond? Delaying strategies and habituation effects, Int. J. Soc. Robotics 1, 141–155 (2009)CrossRefGoogle Scholar
  177. 72.177
    M.P. Michalowski, S. Sabanovic, R. Simmons: A spatial model of engagement for a social robot, IEEE Int. Workshop Adv. Motion Control (2006) pp. 762–767Google Scholar
  178. 72.178
    K. Dautenhahn, M.L. Walters, S. Woods, K.L. Koay, C.L. Nehaniv, E.A. Sisbot, R. Alami, T. Siméon: How may i serve you? A robot companion approaching a seated person in a helping context, ACM/IEEE Int. Conf. Hum.-Robots Interact. (HRI) (2006) pp. 172–179Google Scholar
  179. 72.179
    S. Satake, T. Kanda, D.F. Glas, M. Imai, H. Ishiguro, N. Hagita: How to approach humans? Strategies for social robots to initiate interaction, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2009) pp. 109–116Google Scholar
  180. 72.180
    M.L. Walters, K. Dautenhahn, R.T. Boekhorst, K.L. Koay, C. Kaouri, S. Woods, C. Nehaniv, D. Lee, I. Werry: The influence of subjects’ personality traits on personal spatial zones in a human-robot interaction experiment, IEEE Int. Workshop Robot Hum. Interact. Commun. (ROMAN) (2005) pp. 347–352Google Scholar
  181. 72.181
    H. Hüttenrauch, K.S. Eklundh, A. Green, E.A. Topp: Investigating spatial relationships in human-robot interactions, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2006) pp. 5052–5059Google Scholar
  182. 72.182
    D. Feil-Seifer: Distance-based computational models for facilitating robot interaction with children, J. Hum.-Robot Interact. 1, 55–77 (2012)CrossRefGoogle Scholar
  183. 72.183
    F. Yamaoka, T. Kanda, H. Ishiguro, N. Hagita: Developing a model of robot behavior to identify and appropriately respond to implicit attention-shifting, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2009) pp. 133–140Google Scholar
  184. 72.184
    C. Shi, M. Shimada, T. Kanda, H. Ishiguro, N. Hagita: Spatial formation model for initiating conversation, Robotics Sci. Syst. Conf. (RSS) (2011)Google Scholar
  185. 72.185
    C. Torrey, A. Powers, M. Marge, S.R. Fussell, S. Kiesler: Effects of adaptive robot dialogue on information exchange and social relations, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2006) pp. 126–133Google Scholar
  186. 72.186
    Y. Morales, S. Satake, T. Kanda, N. Hagita: Modeling environments from a route perspective, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2011) pp. 441–448Google Scholar
  187. 72.187
    T. Kollar, S. Tellex, D. Roy, N. Roy: Toward understanding natural language directions, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2010) pp. 259–266Google Scholar
  188. 72.188
    T. Matsumoto, S. Satake, T. Kanda, M. Imai, N. Hagita: Do you remember that shop? – Computational model of spatial memory for shopping companion robots, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2012) pp. 447–454Google Scholar
  189. 72.189
    T. Kanda, M. Shiomi, Z. Miyashita, H. Ishiguro, N. Hagita: A communication robot in a shopping mall, IEEE Trans. Robotics 26(5), 897–913 (2010)CrossRefGoogle Scholar
  190. 72.190
    Y. Fernaeus, M. Håkansson, M. Jacobsson, S. Ljungblad: How do you play with a robotic toy animal?: A long-term study of Pleo, Proc. 9th Int. Conf. Interact. Des. Child. (IDC) (ACM, New York, USA 2010) pp. 39–48CrossRefGoogle Scholar
  191. 72.191
    K. Dautenhahn: Socially intelligent robots: Dimensions of human–robot interaction, Philos. Trans. R. Soc. B Biol, Sci. 362(1480), 679–704 (2007)Google Scholar
  192. 72.192
    COGNIRON, FP6 project, coordinated by LAAS:
  193. 72.193
    LIREC, FP7 project, coordinator Queen Mary University of London:
  194. 72.194
    Companions, FP6 project, coordinated by University of Teesside:
  195. 72.195
    CompaniAble, FP7 project, coordinated by University of Reading:
  196. 72.196
  197. 72.197
    K. Dautenhahn, S. Woods, C. Kaouri, M. Walters, K.L. Koay, I. Werry: What is a robot companion - Friend, assistant or butler?, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), Edmonton (2005) pp. 1488–1493Google Scholar
  198. 72.198
    S. Turkle: Authenticity in the age of digital companions, Interact. Stud. 8(3), 501–517 (2007)CrossRefGoogle Scholar
  199. 72.199
    S. Turkle, W. Taggart, C.D. Kidd, O. Daste: Relational artifacts with children and elders: The complexities of cybercompanionship, Connect. Sci. 18(4), 347–361 (2006)CrossRefGoogle Scholar
  200. 72.200
    M. Shiomi, T. Kanda, H. Ishiguro, N. Hagita: Interactive humanoid robots for a science museum, IEEE Intell. Syst. 22, 25–32 (2007)CrossRefGoogle Scholar
  201. 72.201
    J.-Y. Sung, L. Guo, R.E. Grinter, H.I. Christensen: My Roomba Is Rambo: Intimate home appliances, Lect. Notes Comput. Sci. 4717, 145–162 (2007)CrossRefGoogle Scholar
  202. 72.202
    H. Hüttenrauch, K.S. Eklundh: Fetch-and-carry with Cero: Observations from a long-term user study with a service robot, IEEE Int. Workshop Robot Hum. Interact. Commun. (ROMAN) (2002) pp. 158–163CrossRefGoogle Scholar
  203. 72.203
    S. Šabanović, M.P. Michalowski, R. Simmons: Robots in the wild: Observing human-robot social interaction outside the lab, Proceedings AMC 2006 (2006) pp. 576–581Google Scholar
  204. 72.204
    D. Heylen, B. van Dijk, A. Nijholt: Robotic rabbit companions: Amusing or a nuisance?, J. Multimodal User Interfaces 5, 53–59 (2012)CrossRefGoogle Scholar
  205. 72.205
    T. Salter, F. Michaud, H. Larouche: How wild is wild? A taxonomy to categorize the wildness of child-robot interaction, Int. J. Soc. Robotics 2(4), 405–415 (2010)CrossRefGoogle Scholar
  206. 72.206
    T. Kanda, R. Sato, N. Saiwaki, H. Ishiguro: A two-month field trial in an elementary school for long-term human-robot interaction, IEEE Trans. Robotics 23(5), 962–971 (2007)CrossRefGoogle Scholar
  207. 72.207
    H. Hűttenrauch, E.A. Topp, E.K. Severinson: The art of gate-crashing – Bringing HRI into users’ homes, Interact. Stud. 10(3), 274–297 (2009)CrossRefGoogle Scholar
  208. 72.208
    T. Bickmore, R. Picard: Establishing and maintaining long-term human-computer relationships, ACM Trans. Comput. Hum. Interact. 59(1), 21–30 (2005)Google Scholar
  209. 72.209
    T. Bickmore, L. Caruso, K. Clough-Gorr, T. Heeren: It’s just like you talk to a friend – Relational agents for older adults, Interact. Comput. 17(6), 711–735 (2005)CrossRefGoogle Scholar
  210. 72.210
    K.L. Koay, D.S. Syrdal, M.L. Walters, K. Dautenhahn: Five weeks in the robot house - exploratory human-robot interaction trials in a domestic setting, IEEE 2nd Int. Conf. Adv. Comput.-Hum. Interact. (ACHI) (2008) pp. 219–226Google Scholar
  211. 72.211
    M.E. Pollack, L. Brown, D. Colbry, C.E. McCarthy, C. Orosz, B. Peintner, I. Tsamardinos: Autominder: An intelligent cognitive orthotic system for people with memory impairment, Robotics Auton. Syst. 44(3), 273–282 (2003)CrossRefGoogle Scholar
  212. 72.212
    A. Cesta, F. Pecora: The robocare project: Intelligent systems for elder care, AAAI Fall Symp. Caring Mach. AI Elder Care, USA (2005)Google Scholar
  213. 72.213
    R. Cuijpers, M. Bruna, J. Ham, E. Torta: Attitude towards robots depends on interaction but not on anticipatory behaviour, Lect. Notes Comput. Sci. 7072, 163–172 (2011)CrossRefGoogle Scholar
  214. 72.214
    C. Huijnen, A. Badii, H. van den Heuvel, P. Caleb-Solly, D. Thiemert: Maybe it becomes a buddy, but do not call it a robot – Seamless cooperation between companion robotics and smart homes, Lect. Notes Comput. Sci. 7040, 324–329 (2011)CrossRefGoogle Scholar
  215. 72.215
    Florence, FP7 European project, coordinated by Philips Electronics Nederland B.V.:
  216. 72.216
    SRS, FP7 European project, coordinated by Cardiff University:
  217. 72.217
    ACCOMPANY, FP7 European project, coordinated by University of Hertfordshire:
  218. 72.218
    Quality of Life Technology Center:
  219. 72.219
    CASALA, Dundalk Institute of Technology:
  220. 72.220
  221. 72.221
    B.J. Fogg: Persuasive computers: Perspectives and research directions, Proc. ACM/SIGCHI Conf. Hum. Factors Comput. Syst., ed. by C.-M. Karat, A. Lund, J. Coutaz, J. Karat (1998) pp. 225–232Google Scholar
  222. 72.222
    B. Mutlu, T. Kanda, J. Forlizzi, J. Hodgins, H. Ishiguro: Conversational gaze mechanisms for humanlike robots, ACM Trans, Interact. Intell. Syst. 1(2), 33 (2012)Google Scholar
  223. 72.223
    M. Salem, S. Kopp, I. Wachsmuth, K. Rohlfing, F. Joublin: Generation and evaluation of communicative robot gesture, Int. J. Soc. Robotics 4(2), 201–217 (2012)CrossRefGoogle Scholar
  224. 72.224
    J. Mumm, B. Mutlu: Human-robot proxemics: Physical and psychological distancing in human-robot interaction, ACM/IEEE 6th Int. Conf. Hum.-robot Interact. (HRI) (2011) pp. 331–338Google Scholar
  225. 72.225
    F. Yamaoka, T. Kanda, H. Ishiguro, N. Hagita: How contingent should a lifelike robot be? The Relationship between contingency and complexity, Connect. Sci. 19(2), 143–162 (2007)CrossRefGoogle Scholar
  226. 72.226
    M. Shimada, T. Kanda: What is the appropriate speech rate for a communication robot?, Interact. Stud. 13(3), 408–435 (2012)CrossRefGoogle Scholar
  227. 72.227
    D. François, D. Polani, K. Dautenhahn: Towards socially adaptive robots: A novel method for real time recognition of human-robot interaction styles, Proc. Humanoids 2008, Daejeon (2008) pp. 353–359Google Scholar
  228. 72.228
    D. Szafir, B. Mutlu: Pay attention! Designing adaptive agents that monitor and improve user engagement, Proc. 30th ACM/SIGCHI Conf. Hum. Factors Comput. (2012)Google Scholar
  229. 72.229
    R. Rose, M. Scheutz, P. Schermerhorn: Towards a conceptual and methodological framework for determining robot believability, Interact. Stud. 11(2), 314–335 (2010)CrossRefGoogle Scholar
  230. 72.230
    J. Forlizzi, C. DiSalvo: Service robots in the domestic environment: A study of the roomba vacuum in the home, Proc. 1st ACM SIGCHI/SIGART Conf. Hum.-Robot Interact. (2006) pp. 258–265CrossRefGoogle Scholar
  231. 72.231
    N.E. Sharkey, A.J.C. Sharkey: The crying shame of robot nannies: An ethical appraisal, J. Interact. Stud. 11, 161–190 (2010)CrossRefGoogle Scholar
  232. 72.232
    A. Sharkey, N. Sharkey: Granny and the robots: Ethical issues in robot care for the elderly, Ethics Inform. Technol. 14(1), 27–40 (2012)CrossRefGoogle Scholar
  233. 72.233
    D. Robert, C. Breazeal: Blended reality characters, Proc. Seventh Annu. ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2012) pp. 359–366CrossRefGoogle Scholar
  234. 72.234
    M. Imai, T. Ono, T. Etani: Agent migration: Communications between a human and robot, IEEE Int. Conf. Syst. Man Cybern., Vol. 4 (1999) pp. 1044–1048Google Scholar
  235. 72.235
    D.S. Syrdal, K.L. Koay, M.L. Walters, K. Dautenhahn: The boy-robot should bark! – Children’s impressions of agent migration into diverse embodiments, Proc. New Front. Hum.-Robot Interact. Symp. AISB Convention (2009) pp. 116–121Google Scholar
  236. 72.236
    K.L. Koay, D.S. Syrdal, M.L. Walters, K. Dautenhahn: A user study on visualization of agent migration between two companion robots, 13th Int. Conf. Hum.-Comput. Interact. (HCII) (2009)Google Scholar
  237. 72.237
    E.M. Segura, H. Cramer, P.F. Gomes, S. Nylander, A. Paiva: Revive!: Reactions to migration between different embodiments when playing with robotic pets, Proc. 11th Int. Conf. Interact. Des. Child. (IDC) (2012) pp. 88–97Google Scholar
  238. 72.238
    M. Kriegel, R. Aylett, P. Cuba, V.M.A. Paiva: Robots meet IVAs: A mind-body interface for migrating artificial intelligent agents, Proc. Intell. Virtual Agents, Reykjavik (2011)Google Scholar
  239. 72.239
    K.L. Koay, D.S. Syrdal, K. Dautenhahn, K. Arent, L. Malek, B. Kreczmer: Companion migration – Initial participants’ feedback from a video-based prototyping study. In: Mixed Reality and Human-Robot Interact, ed. by X. Wang (Springer, Berlin, Heidelberg 2011) pp. 133–151CrossRefGoogle Scholar
  240. 72.240
    W.C. Ho, M. Lim, P.A. Vargas, S. Enz, K. Dautenhahn, R. Aylett: An initial memory model for virtual and robot companions supporting migration and long-term interaction, 18th IEEE Int. Symp. Robot Hum. Interact. Commun. (ROMAN) (2009)Google Scholar
  241. 72.241
    G.M.P. O’Hare, B.R. Duffy, J.F. Bradley, A.N. Martin: Agent chameleons: Moving minds from robots to digital information spaces, Proc. Auton. Minirobots Res. Edutainment (2003) pp. 18–21Google Scholar
  242. 72.242
    Y. Sumi, K. Mase: AgentSalon: Facilitating face-to-face knowledge exchange through conversations among personal agents, Proc. 5th Int. Conf. Auton. Agents (AGENTS) (2001) pp. 393–400CrossRefGoogle Scholar
  243. 72.243
    P.K. Allen, A. Miller, P.Y. Oh, B.B. Leibowitz: Integration of vision, force and tactile sensing for grasping, Int. J. Intell. Mach. 4(1), 129–149 (1999)Google Scholar
  244. 72.244
    M. Siegel, M.C. Breazeal, M. Norton: Persuasive robotics: The influence of robot gender on human behavior, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2009) pp. 2563–2568Google Scholar
  245. 72.245
    T. Chen, C.-H. King, A. Thomaz, C. Kemp: Touched by a robot: An investigation of subjective responses to robot-initiated touch, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2011)Google Scholar
  246. 72.246
    P.K. Davis: The Power of Touch – The Basis for Survival, Health, Intimacy, and Emotional Well-Being (Hay House, Carlsbad 1999)Google Scholar
  247. 72.247
    M.J. Hertenstein, J.M. Verkamp, A.M. Kerestes, R.M. Holmes: The communicative functions of touch in humans, non-human primates, and rats: A review and synthesis of the empirical research, Genet. Soc. Gen. Psychol. Monogr. 132(1), 5–94 (2006)CrossRefGoogle Scholar
  248. 72.248
    RoboskinprojectGoogle Scholar
  249. 72.249
    A. Schmitz, P. Maiolino, M. Maggiali, L. Natale, G. Cannata, G. Metta: Methods and technologies for the implementation of large-scale robot tactile sensors, IEEE Trans. Robotics 27(3), 389–400 (2011)CrossRefGoogle Scholar
  250. 72.250
    R.S. Dahiya, M. Getta, M. Valle, G. Sandini: Tactile sensing – From humans to humanoids, IEEE Trans. Robotics 26(1), 1–20 (2010)CrossRefGoogle Scholar
  251. 72.251
    W. Stiehl, L. Lalla, C. Breazeal: A somatic alphabet approach to sensitive skin for robots, Proc. IEEE Int. Conf. Robotics Autom. (ICRA), New Orleans (2004) pp. 2865–2870Google Scholar
  252. 72.252
    W. Stiehl, C. Breazeal: Design of a therapeutic robotic companion for relational, affective touch, Proc. 14th IEEE Workshop Robot Hum. Interact. Commun. (ROMAN), Nashville (2005) pp. 408–415Google Scholar
  253. 72.253
    W.D. Stiehl, C. Breazeal: A sensitive skin for robotic companions featuring temperature, force and electric field sensors, Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2006) pp. 1952–1959Google Scholar
  254. 72.254
    B.D. Argall, A. Billard: A survey of tactile human-robot interactions, Robotics Auton. Syst. 58(10), 1159–1176 (2010)CrossRefGoogle Scholar
  255. 72.255
    J.K. Lee, R.L. Toscano, W.D. Stiehl, C. Breazeal: The design of a semi-autonomous robot avatar for family communication and education, Proc. 17th IEEE Int. Symp. Robot Hum. Interact. Commun. (ROMAN) (2008) pp. 166–173Google Scholar
  256. 72.256
    F. Mueller, F. Vetere, M. Gibbs, J. Kjeldskov, S. Pedell, S. Howard: Hug over a distance, Proc. Conf. Hum. Factors Comput. Syst. (2005) pp. 1673–1676Google Scholar
  257. 72.257
    J.K.S. The, A.D. Cheok, R.L. Peiris, Y. Choi, V. Thuong, S. Lai: Huggy Pajama: A mobile parent and child hugging communication system, Proc. 7th Int. Conf. Interact. Des. Child. (IDC) (2008) pp. 250–257Google Scholar
  258. 72.258
    F. Papadopoulos, K. Dautenhahn, W.C. Ho: Exploring the use of robots as social mediators in a remote human-human collaborative communication experiment, Paladyn 3(1), 1–10 (2012)Google Scholar
  259. 72.259
    B.D. Argall, E. Sauser, A. Billard: Tactile guidance for policy adaptation, Found. Trends Robotics 1(2), 79–133 (2011)CrossRefGoogle Scholar
  260. 72.260
    T. Noda, T. Miyashita, H. Ishiguro, N. Hagita: Super-flexible skin sensors embedded on the whole body, self-organizing based on haptic interactions, Robotics Sci. Syst. Conf. (2008)Google Scholar
  261. 72.261
    T. Tajika, T. Miyashita, H. Ishiguro, N. Hagita: Automatic categorization of haptic interactions-what are the typical haptic interactions between a human and a robot?, Proc. 6th IEEE-RAS Int. Conf. Humanoid Robots (Humanoids), Genova (2006)Google Scholar
  262. 72.262
    H. Knight, R. Toscano, W.D. Stiehl, A. Chang, Y. Wang, C. Breazeal: Real-time social touch gesture recognition for sensate robots, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2009) pp. 3715–3720Google Scholar
  263. 72.263
    S. Yohanan, K.E. MacLean: The role of affective touch in human-robot interaction: Human intent and expectations in touching the haptic creature, Int. J. Soc. Robotics 4, 163–180 (2011)CrossRefGoogle Scholar
  264. 72.264
    W.D. Stiehl, C. Breazeal: Affective touch for robotic companions, Proc. Affect. Comput. Intell. Interact., Bejing (2005)Google Scholar
  265. 72.265
    D. François, K. Dautenhahn, D. Polani: Using real-time recognition of human-robot interaction styles for creating adaptive robot behaviour in robot-assisted play, Proc. 2nd IEEE Symp. Artif. Life, Nashville (2009) pp. 45–52Google Scholar
  266. 72.266
    Aurora project, University of Hertfordshire:
  267. 72.267
    K. Dautenhahn: Robots as social actors: AURORA and the case of autism, Proc. 3rd Int. Cogn. Technol. Conf., San Francisco (1999)Google Scholar
  268. 72.268
    American Psychiatric Association : Diagnostic and Statistical Manual of Mental Disorders DSM-IV (APA, Washington 1995)Google Scholar
  269. 72.269
    R. Jordan: Autistic Spectrum Disorders – An Introductory Handbook for Practitioners (David Fulton, London 1999)Google Scholar
  270. 72.270
    K. Dautenhahn, I. Werry: Towards interactive robots in autism therapy: Background, motivation and challenges, Pragmat. Cogn. 12(1), 1–35 (2004)CrossRefGoogle Scholar
  271. 72.271
    I. Werry, K. Dautenhahn, B. Ogden, W. Harwin: Can social interaction skills be taught by a social agent? The role of a robotic mediator in autism therapy, Proc. 4th Int. Conf. Cogn. Technol. Instrum. Mind (CT), ed. by M. Beynon, C.L. Nehaniv, K. Dautenhahn (Springer, London 2001) pp. 57–74CrossRefGoogle Scholar
  272. 72.272
    B. Scassellati, H. Admoni, M. Mataric: Robots for use in autism research, Annu. Rev. Biomed. Eng. 14, 275–294 (2012)CrossRefGoogle Scholar
  273. 72.273
    I. Werry, K. Dautenhahn, W. Harwin: Evaluating the response of children with autism to a robot, Proc. RESNA Annu. Conf. Rehabil. Eng. Assist. Technol. Soc. N. Am., Nevada (2001)Google Scholar
  274. 72.274
    A. Tapus, A. Peca, A. Aly, C. Pop, L. Jisa, S. Pintea, A.S. Rusi, D.O. David: Children with autism social engagement in interaction with Nao, an imitative robot – A series of single case experiments, Interact. Stud. 13(3), 315–347 (2012)CrossRefGoogle Scholar
  275. 72.275
    B. Vanderborght, R. Simut, J. Saldien, C. Pop, A.S. Rusu, S. Pintea, D. Lefeber, D.O. David: Using the social robot probo as a social story telling agent for children with ASD, Interact. Stud. 13(3), 348–372 (2012)CrossRefGoogle Scholar
  276. 72.276
    A. Billard, B. Robins, K. Dautenhahn, J. Nadel: Building robota, a mini-humanoid robot for the rehabilitation of children with autism, RESNA Assist. Technol. J. 19(1), 37–49 (2006)CrossRefGoogle Scholar
  277. 72.277
    B. Robins, K. Dautenhahn, R. te Boekhorst, A. Billard: Robotic assistants in therapy and education of children with autism: Can a small humanoid robot help encourage social interaction skills?, Univ. Access Inform. Soc. 4(2), 105–120 (2005)CrossRefGoogle Scholar
  278. 72.278
    H. Kozima, M.P. Michalowski, C. Nakagawa: Keepon: A playful robot for research, therapy, and entertainment, Int. J. Soc. Robotics 1(1), 3–18 (2009)CrossRefGoogle Scholar
  279. 72.279
    D. François, S. Powell, K. Dautenhahn: A long-term study of children with autism playing with a robotic pet: Taking inspirations from non-directive play therapy to encourage children’s proactivity and initiative-taking, Interact. Stud. 10(3), 324–373 (2009)CrossRefGoogle Scholar
  280. 72.280
    A. Duquette, F. Michaud, H. Mercier: Exploring the use of a mobile robot as an imitation agent with children with low-functioning autism, Auton. Robots 24, 147–157 (2008)CrossRefGoogle Scholar
  281. 72.281
    J.J. Diehl, L.M. Schmitt, M. Villano, C.R. Crowell: The clinical use of robots for individuals with autism spectrum disorders: A critical review, Res. Autism Spectr. Disord. 6, 249–262 (2012)CrossRefGoogle Scholar
  282. 72.282
    KASPAR, University of Hertfordshire:
  283. 72.283
    B. Robins, K. Dautenhahn: Developing play scenarios for tactile interaction with a humanoid robot: A case study exploration with children with autism, Lect. Notes Comput. Sci. 6414, 243–252 (2010)CrossRefGoogle Scholar
  284. 72.284
    B. Robins, E. Ferrari, K. Dautenhahn, G. Kronrief, B. Prazak, G.J. Gerderblom, F. Caprino, E. Laudanna, P. Marti: Human-centred design methods: Developing scenarios for robot assisted play informed by user panels and field trials, Int. J. Hum.-Comput. Stud. 68, 873–898 (2010)CrossRefGoogle Scholar
  285. 72.285
    J. Wainer, K. Dautenhahn, B. Robins, F. Amirabdollahian: Collaborating with Kaspar: Using an autonomous humanoid robot to foster cooperative dyadic play among children with autism, Proc. IEEE-RAS Int. Conf. Humanoid Robots (2010) pp. 631–638Google Scholar
  286. 72.286
    B. Robbins, K. Dautenhahn, E. Ferrari, G. Kronreif, B. Prazak-Aram, P. Marti, I. Iacono, G.J. Gelderblom, T. Bernd, F. Caprino, E. Laudanna: Scenarios of robot-assisted play for children with cognitive and physical disabilities, Interact. Stud. 13(2), 189–234 (2012)CrossRefGoogle Scholar
  287. 72.287
    B. Robins, K. Dautenhahn, P. Dickerson: Embodiment and cognitive learning – Can a humanoid robot help children with autism to learn about tactile social behaviour?, Lect. Notes Comput. Sci. 7621, 6675 (2012)Google Scholar
  288. 72.288
    Z. Ji, F. Amirabdollahian, D. Polani, K. Dautenhahn: Histogram based classification of tactile patterns on periodically distributed skin sensors for a humanoid robot, Proc. 20th IEEE Int. Symp. Robot Hum. Interact. Commun. (ROMAN) (2011) pp. 433–440Google Scholar
  289. 72.289
    B. Robins, F. Amirabdollahian, Z. Ji, K. Dautenhahn: Tactile interaction with a humanoid robot for children with autism: A case study analysis involving user requirements and results of an initial implementation, Proc. 19th IEEE Int. Symp. Robot Hum. Interact. Commun. (ROMAN) (2010) pp. 704–711CrossRefGoogle Scholar
  290. 72.290
    H. Cramer, N. Kemper, A. Amin, B. Wielinga, V. Evers: Give me a hug: The effects of touch and autonomy on people’s responses to embodied social agents, Comput. Animat. Virtual Worlds 20, 437–445 (2009)CrossRefGoogle Scholar
  291. 72.291
    E.S. Kim, L.D. Berkovits, E.P. Bernier, D. Leyzberg, F. Shic, R. Paul, B. Scassellati: Social robots as embedded reinforcers of social behavior in children with autism, J. Autism Dev. Disord. 43(5), 1038–1049 (2012)CrossRefGoogle Scholar
  292. 72.292
    G. Hoffman, C. Breazeal: Cost-based anticipatory action selection for human–robot fluency, IEEE Trans. Robotics 23, 952–961 (2007)CrossRefGoogle Scholar
  293. 72.293
    G. Hoffman, C. Breazeal: Effects of anticipatory perceptual simulation on practiced human-robot tasks, Auton. Robots 28, 403–423 (2010)CrossRefGoogle Scholar
  294. 72.294
    T. Spexard, S. Li, B. Wrede, J. Fritsch, G. Sagerer, O. Booij, Z. Zivkovic, B. Terwijn, B. Kröse: Biron, Where are you? Enabling a robot to learn new places in a real home environment by integrating spoken dialog and visual localization, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2006) pp. 934–940Google Scholar
  295. 72.295
    S. Woods, K. Dautenhahn, C. Kaouri: Is someone watching me? – Consideration of social facilitation effects in human-robot interaction experiments, IEEE Int. Symp. Comput. Intell. Robotics Autom. (CIRA) (2005) pp. 53–60Google Scholar
  296. 72.296
    N. Riether, F. Hegel, B. Wrede, G. Horstmann: Social facilitation with social robots?, 7th ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2012) pp. 41–47Google Scholar
  297. 72.297
    E. Takano, T. Chikaraishi, Y. Matsumoto, Y. Nakamura, H. Ishiguro, K. Sugamoto: Psychological effects on interpersonal communication by bystander android using motions based on human-like needs, IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS) (2009) pp. 3721–3726Google Scholar
  298. 72.298
    M. Shiomi, T. Kanda, S. Koizumi, H. Ishiguro, N. Hagita: Group attention control for communication robots with Wizard of Oz approach, ACM/IEEE Int. Conf. Hum.-Robot Interact. (HRI) (2007) pp. 121–128Google Scholar
  299. 72.299
    M. Shiomi, K. Nohara, T. Kanda, H. Ishiguro, N. Hagita: Estimating group states for interactive humanoid robots, IEEE-RAS Int. Conf. Humanoid Robots (Humanoids) (2009) pp. 318–323Google Scholar
  300. 72.300
    Y. Matsuyama, H. Taniyama, S. Fujie, T. Kobayashi: Framework of communication activation robot participating in multiparty conversation, AAAI Fall Symp. Ser. (2010) pp. 68–73Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2016

Authors and Affiliations

  • Cynthia Breazeal
    • 1
    Email author
  • Kerstin Dautenhahn
    • 2
  • Takayuki Kanda
    • 3
  1. 1.Personal Robots GroupMIT Media LabCambridgeUSA
  2. 2.School of Computer ScienceUniversity of HertfordshireHatfieldUK
  3. 3.Intelligent Robotics and Communication LaboratoriesAdvanced Telecommunications Research (ATR) Institute InternationalKyotoJapan

Personalised recommendations