Skip to main content

Exploiting Symmetry in Human Robot-Assisted Dressing Using Reinforcement Learning

  • Conference paper
  • First Online:
Progress in Artificial Intelligence (EPIA 2021)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 12981))

Included in the following conference series:

Abstract

In this work, we address the problem of symmetry transfer in human-robot collaborative tasks, i.e., how certain actions can be extended to their symmetrical by exploiting symmetries in their execution. We contribute an approach capable of considering the symmetry inherent to a given task, such as the human or robot’s lateral symmetry, abstracting them from the robot’s decision process. We instantiate our approach in a robot-assisted backpack dressing scenario. A two-manipulator Baxter robot assists a human user in sequentially putting on both straps of a backpack. We evaluate the proposed symmetry-transfer approach in two complementary perspectives: the quality of the agent’s learned policy in a simulated environment and the efficiency of the complete system in a real-life scenario with a robotic platform. The results show that our approach allows the extension of the execution of single-side trained collaborative tasks to their symmetrical with no additional training and minimal performance loss.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 109.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 139.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Average life expectancy by country. https://www.worlddata.info/life-expectancy.php#by-world

  2. Barrett, S., Taylor, M., Stone, P.: Transfer learning for reinforcement learning on a physical robot. In: AAMAS Workshop on Adaptive Learning Agents (2010)

    Google Scholar 

  3. Cao, Z., Hidalgo, G., Simon, T., Wei, S., Sheikh, Y.: OpenPose: realtime multi-person 2D pose estimation using part affinity fields. IEEE Trans. Pattern Anal. Mach. Intell. 43(1), 172–186 (2021)

    Article  Google Scholar 

  4. Christiano, P., et al.: Transfer from simulation to real world through learning deep inverse dynamics model. CoRR abs/1610.03518 (2016)

    Google Scholar 

  5. Gao, Y., Chang, H., Demiris, Y.: User modelling for personalised dressing assistance by humanoid robots. In: Proceedings of 2015 IEEE/RSJ International Conference Intelligent Robots and Systems, pp. 1840–1845 (2015)

    Google Scholar 

  6. Jevtić, A., et al.: Personalized robot assistant for support in dressing. IEEE Trans. Cognitive Developmental Syst. 11(3), 363–374 (2019)

    Article  Google Scholar 

  7. Klee, S., Ferreira, B., Silva, R., Costeira, J., Melo, F., Veloso, M.: Personalized assistance for dressing users. In: Proceedings of 7th International Conference Social Robotics (2015)

    Google Scholar 

  8. Maeda, G., Ewerton, M., Lioutikov, R., Amor, H., Peters, J., Neumann, G.: Learning interaction for collaborative tasks with probabilistic movement primitives. In: Proceedings of the 2014 IEEE-RAS International Conference Humanoid Robots, pp. 527–534 (2014)

    Google Scholar 

  9. Mahler, J., Goldberg, K.: Learning deep policies for robot bin picking by simulating robust grasping sequences. In: Proceedings of 1st Annual Conference Robot Learning (2017)

    Google Scholar 

  10. Olson, E.: AprilTag: a robust and and flexible visual fiducial system. In: Proceedings of 2011 IEEE International Conference Robotics and Automation, pp. 3400–3407 (2011)

    Google Scholar 

  11. Puterman, M.: Markov Decision Processes: Discrete Stochastic Dynamic Programming. John Wiley & Sons (2014)

    Google Scholar 

  12. Rahimi, A., Recht, B.: Weighted sums of random kitchen sinks: Replacing minimization with randomization in learning. In: Advances in Neural Information Processing Systems 22, pp. 1313–1320 (2009)

    Google Scholar 

  13. Roh, Y., Heo, G., Whang, S.: A survey on data collection for machine learning: a big data-AI integration perspective. IEEE Trans. Knowl. Data Eng. 33(4), 1328–1347 (2019)

    Article  Google Scholar 

  14. ce Rosenthal, S., Biswas, J., Veloso, M.: An effective personal mobile robot agent through symbiotic human-robot interaction. In: Proceedings of 9th International Conference Autonomous Agents and Multiagent Systems, pp. 915–922 (2010)

    Google Scholar 

  15. Rusu, A., Večerik, M., Rothörl, T., Heess, N., Pascanu, R., Hadsell, R.: Sim-to-real robot learning from pixels with progressive nets. In: Proceedings of 1st Annual Conference Robot Learning (2017)

    Google Scholar 

  16. Silva, R., Faria, M., Melo, F., Veloso, M.: Adaptive indirect control through communication in collaborative human-robot interaction. In: Proceedings of 2017 IEEE/RSJ International Conference Intelligent Robots and Systems, pp. 3617–3622 (2017)

    Google Scholar 

  17. Silva, R., Melo, F., Veloso, M.: Adaptive symbiotic collaboration for targeted complex manipulation tasks. In: Proceedings of 22nd Eur. Conf. Artificial Intelligence (2016)

    Google Scholar 

  18. Sutton, R.: Integrated architectures for learning, planning, and reacting based on approximating dynamic programming. In: Proceedings of 7th International Conference Machine Learning, pp. 216–224 (1990)

    Google Scholar 

  19. Sutton, R., Barto, A.: Reinforcement Learning: An Introduction. MIT Press (2018)

    Google Scholar 

  20. Veloso, M., Biswas, J., Coltin, B., Rosenthal, S.: CoBots: robust symbiotic autonomous mobile service robots. In: Proceedings of 24th International Joint Conference Artificial Intelligence, pp. 4423–4429 (2015)

    Google Scholar 

  21. Watkins, C.: Learning from delayed rewards. Ph.D. thesis, King’s College, Cambridge University (1989)

    Google Scholar 

Download references

Acknowledgements

This work was partially supported by national funds through the Portuguese Fundação para a Ciência e a Tecnologia under project UIDB/50021/2020 (INESC-ID multi-annual funding) and the Carnegie Mellon Portugal Program and its Information and Communications Technologies Institute, under project CMUP-ERI/HCI/0051/2013. R. Silva acknowledges the PhD grant SFRH/BD/113695/ 2015. M. Vasco acknowledges the PhD grant SFRH/BD/139362/2018.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Pedro Ildefonso .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ildefonso, P. et al. (2021). Exploiting Symmetry in Human Robot-Assisted Dressing Using Reinforcement Learning. In: Marreiros, G., Melo, F.S., Lau, N., Lopes Cardoso, H., Reis, L.P. (eds) Progress in Artificial Intelligence. EPIA 2021. Lecture Notes in Computer Science(), vol 12981. Springer, Cham. https://doi.org/10.1007/978-3-030-86230-5_32

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-86230-5_32

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-86229-9

  • Online ISBN: 978-3-030-86230-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics