Skip to main content

The Effectiveness Index Intrinsic Reward for Coordinating Service Robots

  • Chapter
  • First Online:
Distributed Autonomous Robotic Systems

Part of the book series: Springer Proceedings in Advanced Robotics ((SPAR,volume 6))

Abstract

Modern multi-robot service robotics applications often rely on coordination capabilities at multiple levels, from global (system-wide) task allocation and selection, to local (nearby) spatial coordination to avoid collisions. Often, the global methods are considered to be the heart of the multi-robot system, while local methods are tacked on to overcome intermittent, spatially-limited hindrances. We tackle this general assumption. Utilizing the alphabet soup simulator (simulating order picking, made famous by Kiva Systems), we experiment with a set of myopic, local methods for obstacle avoidance. We report on a series of experiments with a reinforcement-learning approach, using the Effectiveness-Index intrinsic reward, to allow robots to learn to select between methods to use when avoiding collisions. We show that allowing the learner to explore the space of parameterized methods results in significant improvements, even compared to the original methods provided by the simulator.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    This is actually not stated explicitly in [13], but is implied by the design, which explicitly leaves path-planning and motion-planning to each robot’s individual controlling agent.

References

  1. Balch, T., Arkin, R.C.: Behavior-based formation control for multirobot teams. IEEE Trans. Robot. Autom. 14(6), 926–939 (1998)

    Article  Google Scholar 

  2. Bergstra, J., Bengio, Y.: Random search for hyper-parameter optimization. J. Mach. Learn. Res. 12, 281–305 (2012)

    MathSciNet  MATH  Google Scholar 

  3. Bouraine, S., Fraichard, T., Azouaoui, O., Salhi, H.: Passively safe partial motion planning for mobile robots with limited field-of-views in unknown dynamic environments. In: Proceedings of the IEEE International Conference on Robotics and Automation (2014)

    Google Scholar 

  4. Claus, C., Boutilier, C.: The dynamics of reinforcement learning in cooperative multiagent systems. In: Proceedings of the Fifteenth National Conference on Artificial Intelligence (AAAI-98), pp. 746–752 (1998)

    Google Scholar 

  5. Fox, D., Burgard, W., Thrun, S.: The dynamic window approach to collision avoidance. IEEE Robot. Autom. Mag. 4(1), 23–33 (1997)

    Article  Google Scholar 

  6. Godoy, J.E., Karamouzas, I., Guy, S.J., Gini, M.: Adaptive learning for multi-agent navigation. In: Proceedings of the Fourteenth International Joint Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-15), pp. 1577–1585 (2015)

    Google Scholar 

  7. Hazard, C.J., Wurman, P.R.: Alphabet soup: a testbed for studying resource allocation in multi-vehicle systems. In: Proceedings of the 2006 AAAI Workshop on Auction Mechanisms for Robot Coordination, pp. 23–30 (2006)

    Google Scholar 

  8. Kaminka, G.A., Erusalimchik, D., Kraus, S.: Adaptive multi-robot coordination: a game-theoretic perspective. In: Proceedings of IEEE International Conference on Robotics and Automation (ICRA-10) (2010)

    Google Scholar 

  9. Rosenfeld, A., Kaminka, G.A., Kraus, S., Shehory, O.: A study of mechanisms for improving robotic group performance. Artif. Intell. 172(6), 633–655 (2008)

    Article  MATH  Google Scholar 

  10. Rybski, P., Larson, A., Lindahl, M., Gini, M.: Performance evaluation of multiple robots in a search and retrieval task. In: Proceedings of the Workshop on Artificial Intelligence and Manufacturing, pp. 153–160. Albuquerque, NM (1998)

    Google Scholar 

  11. van den Berg, J., Guy, S., Lin, M., Manocha, D.: Reciprocal n-body collision avoidance. In: Robotics Research pp. 3–19 (2011)

    Google Scholar 

  12. Vaughan, R., Støy, K., Sukhatme, G., Matarić, M.: Go ahead, make my day: robot conflict resolution by aggressive competition. In: Proceedings of the 6th International Conference on the Simulation of Adaptive Behavior. Paris, France (2000)

    Google Scholar 

  13. Wurman, P.R., DÁndrea, R., Mountz, M.: Coordinating hundreds of cooperative, autonomous vehicles in warehouses. AI Mag. (2008)

    Google Scholar 

Download references

Acknowledgements

We gratefully acknowledge support by ISF grants #1511/12, and #1865/16, and good advice from Avi Seifert. As always, thanks to K. Ushi.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Gal A. Kaminka .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer International Publishing AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Douchan, Y., Kaminka, G.A. (2018). The Effectiveness Index Intrinsic Reward for Coordinating Service Robots. In: Groß, R., et al. Distributed Autonomous Robotic Systems. Springer Proceedings in Advanced Robotics, vol 6. Springer, Cham. https://doi.org/10.1007/978-3-319-73008-0_21

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-73008-0_21

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-73006-6

  • Online ISBN: 978-3-319-73008-0

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics