Abstract
The motion planning problem for multi-agent systems becomes particularly challenging when humans or human-controlled robots are present in a mixed environment. To address this challenge, this paper presents an interaction-aware motion planning approach based on game theory in a receding-horizon manner. Leveraging the framework provided by dynamic potential games for handling the interactions among agents, this approach formulates the multi-agent motion planning problem as a differential potential game, highlighting the effectiveness of constrained potential games in facilitating interactive motion planning among agents. Furthermore, online learning techniques are incorporated to dynamically learn the unknown preferences and models of humans or human-controlled robots through the analysis of observed data. To evaluate the effectiveness of the proposed approach, numerical simulations are conducted, demonstrating its capability to generate interactive trajectories for all agents, including humans and human-controlled agents, operating within the mixed environment. The simulation results illustrate the effectiveness of the proposed approach in handling the complexities of multi-agent motion planning in real-world scenarios.
Similar content being viewed by others
References
Yarlagadda, R. T. (2015). Future of robots, AI and automation in the United States. IEJRD-International Multidisciplinary Journal, 1(5), 6.
Goel, R., & Gupta, P. (2020). Robotics and industry 4.0. A Roadmap to Industry 4.0: Smart Production, Sharp Business and Sustainable Development, pp. 157–169, Cham: Springer.
Sheridan, T. B. (2016). Human-robot interaction: Status and challenges. Human Factors, 58(4), 525–532.
Valtazanos, A., & Ramamoorthy, S. (2013). Bayesian interaction shaping: Learning to influence strategic interactions in mixed robotic domains. AAMAS, 13, 63–70.
Kolbeinsson, A., Lagerstedt, E., & Lindblom, J. (2019). Foundation for a classification of collaboration levels for human-robot cooperation in manufacturing. Production and Manufacturing Research, 7(1), 448–471.
Di, X., & Shi, R. (2021). A survey on autonomous vehicle control in the era of mixed-autonomy: From physics-based to AI-guided driving policy learning. Transportation Research Part C: Emerging Technologies, 125, 103008.
Leung, K., Schmerling, E., Chen, M., Talbot, J., Gerdes, J. C., & Pavone, M. (2020). On infusing reachability-based safety assurance within probabilistic planning frameworks for human-robot vehicle interactions. In Proceedings of the 2018 International Symposium on Experimental Robotics (pp. 561–574). Springer
Jin, Z., & Pagilla, P. R. (2020). Collaborative operation of robotic manipulators with human intent prediction and shared control. In IEEE International Conference on Human-Machine Systems (ICHMS) (pp. 1–6). IEEE.
Sadigh, D., Landolfi, N., Sastry, S. S., Seshia, S. A., & Dragan, A. D. (2018). Planning for cars that coordinate with people: Leveraging effects on human actions for planning and active information gathering over human internal state. Autonomous Robots, 42, 1405–1426.
Kavuncu, T., Yaraneri, A., & Mehr, N. (2021). Potential iLQR: A potential-minimizing controller for planning multi-agent interactive trajectories. arXiv:2107.04926
Liniger, A., & Lygeros, J. (2019). A noncooperative game approach to autonomous racing. IEEE Transactions on Control Systems Technology, 28(3), 884–897.
Facchinei, F., & Kanzow, C. (2010). Generalized Nash equilibrium problems. Annals of Operations Research, 175(1), 177–211.
Zazo, S., Macua, S. V., Sánchez-Fernández, M., & Zazo, J. (2016). Dynamic potential games with constraints: Fundamentals and applications in communications. IEEE Transactions on Signal Processing, 64(14), 3806–3821.
Swenson, B., Murray, R., & Kar, S. (2018). On best-response dynamics in potential games. SIAM Journal on Control and Optimization, 56(4), 2734–2767.
Marden, J. R., Arslan, G., & Shamma, J. S. (2009). Cooperative control and potential games. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 39(6), 1393–1407.
Williams, Z., Chen, J., & Mehr, N. (2023). Distributed potential iLQR: Scalable game-theoretic trajectory planning for multi-agent interactions. arXiv:2303.04842
Bhatt, M., Yaraneri, A., & Mehr, N. (2022). Efficient constrained multi-agent interactive planning using constrained dynamic potential games. arXiv:2206.08963.
Monderer, D., & Shapley, L. S. (1996). Potential games. Games and Economic Behavior, 14(1), 124–143.
Fonseca-Morales, A., & Hernández-Lerma, O. (2018). Potential differential games. Dynamic Games and Applications, 8, 254–279.
Candogan, O., Menache, I., Ozdaglar, A., & Parrilo, P. A. (2011). Flows and decompositions of games: Harmonic and potential games. Mathematics of Operations Research, 36(3), 474–503.
Hatz, K., Schloder, J. P., & Bock, H. G. (2012). Estimating parameters in optimal control problems. SIAM Journal on Scientific Computing, 34(3), 1707–1728.
Peters, L., Fridovich-Keil, D., Rubies-Royo, V., Tomlin, C. J., & Stachniss, C. (2021). Inferring objectives in continuous dynamic games from noise-corrupted partial state observations. arXiv:2106.03611
Menner, M., Worsnop, P., & Zeilinger, M. N. (2019). Constrained inverse optimal control with application to a human manipulation task. IEEE Transactions on Control Systems Technology, 29(2), 826–834.
Menner, M., & Zeilinger, M. N. (2020). Maximum likelihood methods for inverse learning of optimal controllers. IFAC-PapersOnLine, 53(2), 5266–5272.
Park, T., & Levine, S. (2013). Inverse optimal control for humanoid locomotion. In Robotics Science and Systems Workshop on Inverse Optimal Control and Robotic Learning from Demonstration (pp. 4887–4892).
Evens, B., Schuurmans, M., & Patrinos, P. (2022). Learning MPC for interaction-aware autonomous driving: A game-theoretic approach. In 2022 European Control Conference (ECC) (pp. 34–39). IEEE.
Lin, J., Wang, M., & Wu, H.-N. (2023). Composite adaptive online inverse optimal control approach to human behavior learning. Information Sciences, 638, 118977.
Takei, R., Huang, H., Ding, J., & Tomlin, C. J. (2012). Time-optimal multi-stage motion planning with guaranteed collision avoidance via an open-loop game formulation. In 2012 IEEE International Conference on Robotics and Automation (pp. 323–329). IEEE.
Başar, T., & Olsder, G. J. (1998). Dynamic noncooperative game theory. SIAM.
Sadigh, D., Sastry, S., Seshia, S. A., & Dragan, A. D. (2016). Planning for autonomous cars that leverage effects on human actions. Robotics: Science and Systems, 2, 1–9.
Fridovich-Keil, D., Ratner, E., Peters, L., Dragan, A. D., & Tomlin, C. J. (2020). Efficient iterative linear-quadratic approximations for nonlinear multi-player general-sum differential games. In 2020 IEEE International Conference on Robotics and Automation (ICRA) (pp. 1475–1481). IEEE.
Dreves, A., & Gerdts, M. (2018). A generalized Nash equilibrium approach for optimal control problems of autonomous cars. Optimal Control Applications and Methods, 39(1), 326–342.
Curtis, F. E., & Overton, M. L. (2012). A sequential quadratic programming algorithm for nonconvex, nonsmooth constrained optimization. SIAM Journal on Optimization, 22(2), 474–500.
Lukšan, L., Matonoha, C., & Vlček, J. (2004). Interior-point method for non-linear non-convex optimization. Numerical Linear Algebra with Applications, 11(5–6), 431–453.
Lin, Q., Ma, R., & Xu, Y. (2022). Complexity of an inexact proximal-point penalty method for constrained smooth non-convex optimization. Computational Optimization and Applications, 82(1), 175–224.
Andersson, J., Åkesson, J., & Diehl, M. (2012). CasADi: A symbolic package for automatic differentiation and optimal control. In Recent Advances in Algorithmic Differentiation (pp. 297–307). Springer.
Wächter, A., & Biegler, L. T. (2006). On the implementation of an interior-point filter line-search algorithm for large-scale nonlinear programming. Mathematical Programming, 106, 25–57.
Sabatino, F. (2015). Quadrotor control: Modeling, nonlinear control design, and simulation, Ph.D. Thesis. KTH, School of Electrical Engineering (EES).
Maroger, I., Stasse, O., & Watier, B. (2020). Walking human trajectory models and their application to humanoid robot locomotion. In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 3465–3472). IEEE.
Mombaur, K., Truong, A., & Laumond, J.-P. (2010). From human to humanoid locomotion-an inverse optimal control approach. Autonomous Robots, 28, 369–383.
Author information
Authors and Affiliations
Corresponding author
Additional information
This work was supported by the A\(^\star \)STAR under its “RIE2025 IAF-PP Advanced ROS2-native Platform Technologies for Cross sectorial Robotics Adoption (M21K1a0104)” programme.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Zhang, X., Xie, L. Game-theoretic multi-agent motion planning in a mixed environment. Control Theory Technol. (2024). https://doi.org/10.1007/s11768-024-00207-9
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11768-024-00207-9