Skip to main content
Log in

Teaching Robot’s Proactive Behavior Using Human Assistance

  • Published:
International Journal of Social Robotics Aims and scope Submit manuscript

Abstract

In recent years, there has been a growing interest in enabling autonomous social robots to interact with people. However, many questions remain unresolved regarding the social capabilities robots should have in order to perform this interaction in an ever more natural manner. In this paper, we tackle this problem through a comprehensive study of various topics involved in the interaction between a mobile robot and untrained human volunteers for a variety of tasks. In particular, this work presents a framework that enables the robot to proactively approach people and establish friendly interaction. To this end, we provided the robot with several perception and action skills, such as that of detecting people, planning an approach and communicating the intention to initiate a conversation while expressing an emotional status. We also introduce an interactive learning system that uses the person’s volunteered assistance to incrementally improve the robot’s perception skills. As a proof of concept, we focus on the particular task of online face learning and recognition. We conducted real-life experiments with our Tibi robot to validate the framework during the interaction process. Within this study, several surveys and user studies have been realized to reveal the social acceptability of the robot within the context of different tasks.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17
Fig. 18

Similar content being viewed by others

References

  1. Anzalone SM, Boucenna S, Ivaldi S, Chetouani M (2015) Evaluating the engagement with social robots. Int J Soc Robot 7(4):465–478

    Article  Google Scholar 

  2. Arras K, Mozos O, Burgard W (2007) Using boosted features for the detection of people in 2d range data. In: Proceedings of the IEEE international conference on robotics and automation, pp 3402–3407

  3. Atkinson DJ, Clark MH (2014) Methodology for study of human-robot social interaction in dangerous situations. In: Proceedings of the second international conference on Human-agent interaction. ACM, pp 371–376

  4. Bauer A, Klasing K, Lidoris G, Mühlbauer Q, Rohrmüller F, Sosnowski S, Xu T, Kühnlenz K, Wollherr D, Buss M (2009) The autonomous city explorer: towards natural human-robot interaction in urban environments. Int J Soc Robot 1(2):127–140

    Article  Google Scholar 

  5. Bittermann A, Kuhnlenz K, Buss M (2007) On the evaluation of emotion expressing robots. In: IEEE international conference on robotics and automation, pp 2138–2143

  6. Breazeal C (2003) Emotion and sociable humanoid robots. Int J Hum Comput Stud 59(1):119–155

    Article  Google Scholar 

  7. Breazeal C (2009) Role of expressive behaviour for robots that learn from people. Philos Trans R Soc B Biol Sci 364(1535):3527–3538

    Article  Google Scholar 

  8. Burgard W, Cremers A, Fox D, Hähnel D, Lakemeyer G, Schulz D, Steiner W, Thrun S (1998) The interactive museum tour-guide robot. In: Proceedings of the National Conference on Artificial Intelligence. John Wiley & Sons Ltd, pp 11–18

  9. Clark H (1996) Using language, vol 4. Cambridge University Press, Cambridge

    Book  Google Scholar 

  10. Cooper B, Brna P, Martins A (2000) Effective affective in intelligent systems–building on evidence of empathy in teaching and learning. In: Affective interactions. Springer, Berlin, pp 21–34

  11. Dautenhahn K, Walters M, Woods S, Koay K, Nehaniv C, Sisbot A, Alami R, Siméon T (2006) How may i serve you?: A robot companion approaching a seated person in a helping context. In: Proceedings of the 1st ACM SIGCHI/SIGART conference on Human-robot interaction. ACM, pp 172–179

  12. Dautenhahn K, Woods S, Kaouri C, Walters M, Koay K, Werry I (2005) What is a robot companion-friend, assistant or butler? In: Proceedings of the IEEE/RSJ international conference on intelligent robots and systems, pp 1192–1197

  13. Fellous JM, Arbib MA (2005) Who needs emotions?: The brain meets the robot. Oxford University Press, Oxford

    Book  Google Scholar 

  14. Felzenszwalb PF, Girshick RB, McAllester D, Ramanan D (2010) Object detection with discriminatively trained part-based models. IEEE Trans Pattern Anal Mach Intell 32(9):1627–1645

    Article  Google Scholar 

  15. Ferrer G, Garrell A, Sanfeliu A (2013) Robot companion: a social-force based approach with human awareness-navigation in crowded environments. In: Proceedings of IEEE/RSJ international conference on intelligent robots and systems, pp 1688–1694

  16. Ferrer G, Garrell A, Villamizar M, Huerta I, Sanfeliu A (2013) Robot interactive learning through human assistance. In: Multimodal interaction in image and video applications. Springer, pp 185–203

  17. Garrell A, Sanfeliu A (2012) Cooperative social robots to accompany groups of people. Int J Robot Res 31(13):1675–1701

    Article  Google Scholar 

  18. Garrell A, Sanfeliu A, Moreno-Noguer F (2009) Discrete time motion model for guiding people in urban areas using multiple robots. In Proceedings of the IEEE/RSJ international conference on intelligent robots and systems, pp 486–491

  19. Gill A (1962) Introduction to the theory of finite-state machines. McGraw-Hill, New York

    MATH  Google Scholar 

  20. Gockley R, Forlizzi J, Simmons R (2006) Interactions with a moody robot. In: Proceedings of the 1st ACM SIGCHI/SIGART conference on Human-robot interaction. ACM, pp 186–193

  21. Gockley R, Simmons R, Forlizzi J (2006) Modeling affect in socially interactive robots. In: The 15th IEEE international symposium on robot and human interactive communication. IEEE, pp 558–563

  22. Goffman E (1966) Behavior in public places: notes on the social organization of gatherings. Free Press, New York

  23. Haasch A, Hohenner S, Hüwel S, Kleinehagenbrock M, Lang S, Toptsis I, Fink G, Fritsch J, Wrede B, Sagerer G (2004) Biron–the bielefeld robot companion. In: Proceedings of the international workshop on advances in service robotics. Fraunhofer IRB Verlag, Stuttgart, Germany,pp 27–32

  24. Hall ET (1969) The hidden dimension. Anchor Books, New York

    Google Scholar 

  25. Haralock RM, Shapiro LG (1991) Computer and robot vision. Addison-Wesley Longman Publishing Co., Inc., New York

    Google Scholar 

  26. Hayashi K, Sakamoto D, Kanda T, Shiomi M, Koizumi S, Ishiguro H, Ogasawara T, Hagita N (2007) Humanoid robots as a passive-social medium: a field experiment at a train station. In: Proceedings of the ACM/IEEE international conference on Human-robot interaction, pp 137–144

  27. Helbing D, Molnar P (1995) Social force model for pedestrian dynamics. Phys Rev E 51(5):4282–4286

    Article  Google Scholar 

  28. Hirth J (2012) Towards socially interactive robots-designing an emotion-based control architecture. Dr. Hut

  29. Hirth J, Schmitz N, Berns K (2011) Towards social robots: designing an emotion-based architecture. Int J Soc Robot 3(3):273–290

    Article  Google Scholar 

  30. Höök K (2009) Affective loop experiences: designing for interactional embodiment. Philos Trans R Soc London B Biol Sci 364(1535):3585–3595

    Article  Google Scholar 

  31. Kanda T, Shiomi M, Miyashita Z, Ishiguro H, Hagita N (2009) An affective guide robot in a shopping mall. In: Proceedings of the 4th ACM/IEEE international conference on Human robot interaction, pp 173–180

  32. Kędzierski J, Muszyński R, Zoll C, Oleksy A, Frontkiewicz M (2013) Emys-emotive head of a social robot. Int J Soc Robot 5(2):237–249

    Article  Google Scholar 

  33. Khatib O (1985) Real-time obstacle avoidance for manipulators and mobile robots. Proceedings of the IEEE international conference on robotics and automation, vol 2, pp 500–505

  34. Kirby R (2010) Social robot navigation. Ph.D. thesis, Robotics Institute, Carnegie Mellon University, Pittsburgh, USA

  35. Koay K, Sisbot E, Syrdal D, Walters M, Dautenhahn K, Alami R (2007) Exploratory study of a robot approaching a person in the context of handing over an object. In: Proceedings of the AAAI-SS on multi-disciplinary collaboration for socially assistive robotics, pp 18–24

  36. Kühnlenz B, Sosnowski S, Buß M, Wollherr D, Kühnlenz K, Buss M (2013) Increasing helpfulness towards a robot by emotional adaption to the user. Int J Soc Robot 5(4):457–476

    Article  Google Scholar 

  37. Lee M, Makatchev M (2009) How do people talk with a robot?: An analysis of human-robot dialogues in the real world. In: Proceedings of the international conference extended abstracts on Human factors in computing systems. ACM, pp 3769–3774

  38. Leite I, Martinho C, Pereira A, Paiva A (2009) As time goes by: long-term evaluation of social presence in robotic companions. In: RO-MAN 2009-The 18th IEEE international symposium on robot and human interactive communication, pp 669–674

  39. Ljungblad S, Kotrbova J, Jacobsson M, Cramer H, Niechwiadowicz K (2012) Hospital robot at work: something alien or an intelligent colleague? In: Proceedings of the ACM 2012 conference on computer supported cooperative Work. ACM, pp 177–186

  40. Marsella SC, Gratch J (2009) EMA: a process model of appraisal dynamics. Cogn Syst Res 10(1):70–90

    Article  Google Scholar 

  41. Michalowski M, Sabanovic S, Simmons R (2006) A spatial model of engagement for a social robot. In: Proceedings of the 9th IEEE international workshop on advanced motion control, pp 762–767

  42. Morales Saiki L, Satake S, Huq R, Glass D, Kanda T, Hagita N (2012) How do people walk side-by-side?: using a computational model of human behavior for a social robot. In: Proceedings of the 7th ACM/IEEE international conference on human-robot interaction, pp 301–308

  43. Mutlu B, Shiwa T, Kanda T, Ishiguro H, Hagita N (2009) Footing in human-robot conversations: how robots might shape participant roles using gaze cues. In: Proceedings of the 4th ACM/IEEE international conference on Human robot interaction, pp 61–68

  44. Quigley M, Conley K, Gerkey B, Faust J, Foote T, Leibs J, Ng A (2009) Ros: an open-source robot operating system. In: ICRA workshop on open source software, vol 3, no 3.2, p 5

  45. Rich C, Ponsler B, Holroyd A, Sidner CL (2010) Recognizing engagement in human-robot interaction. In: Human-robot interaction (HRI), 2010 5th ACM/IEEE international conference on. IEEE, pp 375–382

  46. Roseman IJ (1996) Appraisal determinants of emotions: constructing a more accurate and comprehensive theory. Cogn Emot 10(3):241–278

    Article  Google Scholar 

  47. Saldien J, Goris K, Vanderborght B, Vanderfaeillie J, Lefeber D (2010) Expressing emotions with the social robot probo. Int J Soc Robot 2(4):377–389

    Article  Google Scholar 

  48. Salichs MA, Malfaz M (2006) Using emotions on autonomous agents. the role of happiness, sadness and fear. Integr Approaches Mach Conscious Part AISB 6:157–164

    Google Scholar 

  49. Satake S, Kanda T, Glas D, Imai M, Ishiguro H, Hagita N (2009) How to approach humans?-strategies for social robots to initiate interaction. In: Proceedings of the ACM/IEEE international conference on human-robot interaction, pp 109–116

  50. Schlosberg H (1954) Three dimensions of emotion. Psychol Rev 61(2):81

    Article  Google Scholar 

  51. Shi C, Shimada M, Kanda T, Ishiguro H, Hagita N (2011) Spatial formation model for initiating conversation. In: Proceedings of robotics: science and systems, Los Angeles

  52. Shiomi M, Kanda T, Glas D, Satake S, Ishiguro H, Hagita N (2009) Field trial of networked social robots in a shopping mall. In: Proceedings of the IEEE/RSJ international conference on intelligent robots and systems, pp 2846–2853

  53. Sidner CL, Kidd CD, Lee C, Lesh N (2004) Where to look: a study of human-robot engagement. In: Proceedings of the 9th international conference on Intelligent user interfaces. ACM, pp 78–84

  54. Sisbot EA, Marin-Urias LF, Broquere X, Sidobre D, Alami R (2010) Synthesizing robot motions adapted to human presence. Int J Soc Robot 2(3):329–343

    Article  Google Scholar 

  55. Timofte R, Zimmermann K, Van Gool L (2014) Multi-view traffic sign detection, recognition, and 3d localisation. Mach Vis Appl 25(3):633–647

    Article  Google Scholar 

  56. Villamizar M, Garrell A, Sanfeliu A, Moreno-Noguer F (2012) Online human-assisted learning using random ferns. In: Proceedings of the 21st international conference on pattern recognition, pp 2821–2824

  57. Walters M, Dautenhahn K, Woods S, Koay K (2007) Robotic etiquette: results from user studies involving a fetch and carry task. In: Proceedings of the ACM/IEEE international conference on human-robot interaction, pp 317–324

  58. Xu J, Broekens J, Hindriks K, Neerincx MA (2014) Robot mood is contagious: effects of robot body language in the imitation game. In: Proceedings of the 2014 international conference on Autonomous agents and multi-agent systems. International Foundation for Autonomous Agents and Multiagent Systems, pp 973–980

  59. Yohanan S, MacLean KE (2012) The role of affective touch in human-robot interaction: human intent and expectations in touching the haptic creature. Int J Soc Robot 4(2):163–180

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to A. Garrell.

Additional information

This work was supported by the Spanish Ministry of Science and Innovation, project Rob-Int-Coop DPI2013-42458-P and AEROARMS European project H2020-ICT-2014-1-644271.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Garrell, A., Villamizar, M., Moreno-Noguer, F. et al. Teaching Robot’s Proactive Behavior Using Human Assistance. Int J of Soc Robotics 9, 231–249 (2017). https://doi.org/10.1007/s12369-016-0389-0

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12369-016-0389-0

Keywords

Navigation