Abstract
Successful human-machine teaming often hinges on the ability of eXplainable Artificial Intelligence (XAI) to make an agent’s reasoning transparent to human teammates. Doing so requires that the agent navigate a tradeoff between revealing its reasoning to those teammates without overwhelming them with too much information. This challenge is amplified when a person is teamed with multiple agents. This amplification is not simply linear, due to the increase from 1 to N agents’ worth of reasoning content, but also due to the interdependency among the agents’ reasoning that must be made transparent as well. In this work, we examine the challenges in conveying this interdependency to people teaming with multiple agents. We also propose alternate domain-independent strategies for a team of simulated robots to generate messages about their reasoning to be conveyed to a human teammate. We illustrate these strategies through their implementation in a search-and-rescue simulation testbed.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Adams, J.A.: Human-robot interaction design: Understanding user needs and requirements. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, pp. 447–451 (2005)
Bernstein, D.S., Givan, R., Immerman, N., Zilberstein, S.: The complexity of decentralized control of Markov decision processes. Math. Oper. Res. 27(4), 819–840 (2002)
Bernstein, D.S., Hansen, E.A., Zilberstein, S.: Bounded policy iteration for decentralized POMDPs. In: Proceedings of the International Joint Conference on Artificial Intelligence, pp. 52–57 (2005)
Chen, J.Y., Barnes, M.J., Harper-Sciarini, M.: Supervisory control of multiple robots: Human-performance issues and user-interface design. IEEE Trans. Syst. Man Cybernet. Part C (Appli. Rev.) 41(4), 435–454 (2010)
Chen, J.Y., Barnes, M.J., Kenny, C.: Effects of unreliable automation and individual differences on supervisory control of multiple ground robots. In: Proceedings of the International Conference on Human-Robot Interaction, pp. 371–378 (2011)
Chen, J.Y., Barnes, M.J., Qu, Z.: RoboLeader: An agent for supervisory control of multiple robots. In: Proceedings of the International Conference on Human-Robot Interaction, pp. 81–82 (2010)
Chien, S.Y., Lewis, M., Mehrotra, S., Sycara, K.: Imperfect automation in scheduling operator attention on control of multi-robots. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, pp. 1169–1173 (2013)
Cummings, M.L., Bruni, S., Mercier, S., Mitchell, P.: Automation architecture for single operator, multiple UAV command and control. Tech. rep, Massachusetts Institute Of Technology (2007)
Doshi, P., Gmytrasiewicz, P.J.: Monte Carlo sampling methods for approximating interactive POMDPs. J. Artifi. Intell. Res. 34, 297–337 (2009)
Gmytrasiewicz, P.J., Doshi, P.: Interactive POMDPs: properties and preliminary results. In: Proceedings of the International Joint Conference on Autonomous Agents and Multiagent Systems, vol. 3, pp. 1374–1375 (2004)
Hansen, E.A., Bernstein, D.S., Zilberstein, S.: Dynamic programming for partially observable stochastic games. In: AAAI, vol. 4, pp. 709–715 (2004)
Kaelbling, L.P., Littman, M.L., Cassandra, A.R.: Planning and acting in partially observable stochastic domains. Artif. Intell. 101(1–2), 99–134 (1998)
McKinsey: A future that works: Automation, employment, and productivity. Tech. rep., McKinsey Global Institute (2017)
NSF: National artificial intelligence (AI) research institutes: Accelerating research, transforming society, and growing the American workforce. Tech. rep., National Science Foundation (2020)
Oliehoek, F.A., Amato, C.: A concise introduction to decentralized POMDPs. Springer (2016). https://doi.org/10.1007/978-3-319-28929-8
Papadimitriou, C.H., Tsitsiklis, J.N.: The complexity of Markov decision processes. Math. Oper. Res. 12(3), 441–450 (1987)
Pynadath, David V.., Barnes, Michael J.., Wang, Ning, Chen, Jessie Y. C..: Transparency communication for machine learning in human-automation interaction. In: Zhou, Jianlong, Chen, Fang (eds.) Human and Machine Learning. HIS, pp. 75–90. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-90403-0_5
Pynadath, D.V., Gurney, N., Wang, N.: Explainable reinforcement learning in human-robot teams: The impact of decision-tree explanations on transparency. In: IEEE International Conference on Robot and Human Interactive Communication (2022)
Pynadath, D.V., Marsella, S.C.: PsychSim: Modeling theory of mind with decision-theoretic agents. In: Proceedings of the International Joint Conference on Artificial Intelligence, pp. 1181–1186 (2005)
Rosenfeld, A., Agmon, N., Maksimov, O., Kraus, S.: Intelligent agent supporting human-multi-robot team collaboration. Artif. Intell. 252, 211–231 (2017)
Rovira, E., McGarry, K., Parasuraman, R.: Effects of unreliable automation on decision making in command and control. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 46, pp. 428–432 (2002)
Sarter, N.B., Schroeder, B.: Supporting decision making and action selection under time pressure and uncertainty: The case of in-flight icing. Hum. Factors 43(4), 573–583 (2001)
Szer, D., Charpillet, F., Zilberstein, S.: MAA*: A heuristic search algorithm for solving decentralized POMDPs. In: Proceedings of the Conference on Uncertainty in Artificial Intelligence (2005)
Velagapudi, P., Scerri, P.: Scaling human-robot systems. In: Proceedings of the ACM CHI Conference on Human Factors in Computing Systems (2009)
Wang, N., Pynadath, D.V., Hill, S.G.: Building trust in a human-robot team with automatically generated explanations. In: Proceedings of the Interservice/Industry Training, Simulation and Education Conference, vol. 15315, pp. 1–12 (2015)
Wang, N., Pynadath, D.V., Hill, S.G.: The impact of POMDP-generated explanations on trust and performance in human-robot teams. In: Proceedings of the International Joint Conference on Autonomous Agents and Multi-Agent Systems, pp. 997–1005 (2016)
Zigoris, P., Siu, J., Wang, O., Hayes, A.T.: Balancing automated behavior and human control in multi-agent systems: A case study in RoboFlag. In: Proceedings of the American Control Conference, vol. 1, pp. 667–671 (2003)
Acknowledgments
This work was sponsored by the U.S. Army Research Laboratory (ARL) under contract number W911NF-14-D-0005 and the Defense Advanced Research Projects Agency (DARPA) under contract number W911NF2010011. Statements and opinions expressed do not necessarily reflect the position or the policy of the United States Government or the Defense Advanced Research Projects Agency, and no official endorsements should be inferred.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Wang, N., Pynadath, D.V., Gurney, N. (2023). The Design of Transparency Communication for Human-Multirobot Teams. In: Degen, H., Ntoa, S. (eds) Artificial Intelligence in HCI. HCII 2023. Lecture Notes in Computer Science(), vol 14051. Springer, Cham. https://doi.org/10.1007/978-3-031-35894-4_23
Download citation
DOI: https://doi.org/10.1007/978-3-031-35894-4_23
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-35893-7
Online ISBN: 978-3-031-35894-4
eBook Packages: Computer ScienceComputer Science (R0)