Abstract
We consider the generalized flocking problem in multiagent systems, where the agents must drive a subset of their state variables to common values, while communication is constrained by a proximity relationship in terms of another subset of variables. We build a flocking method for general nonlinear agent dynamics, by using at each agent a near-optimal control technique from artificial intelligence called optimistic planning. By defining the rewards to be optimized in a well-chosen way, the preservation of the interconnection topology is guaranteed, under a controllability assumption. We also give a practical variant of the algorithm that does not require to know the details of this assumption, and show that it works well in experiments on nonlinear agents.
Similar content being viewed by others
References
R. Olfati-Saber, J. A. Fax, R. M. Murray. Consensus and cooperation in networked multiagent systems. Proceedings of the IEEE, 2007, 95(1): 215–233.
W. Ren, R. W. Beard. Distributed Consensus in Multi-vehicle Cooperative Control: Theory and Applications. Communications and Control Engineering. Berlin: Springer, 2008.
R. Olfati-Saber. Flocking for multi-agent dynamic systems: Algorithms and theory. IEEE Transactions on Automatic Control, 2006, 51(3): 401–420.
H. G. Tanner, A. Jadbabaie, G. J. Pappas. Flocking in fixed and switching networks. IEEE Transactions on Automatic Control, 2007, 52(5): 863–868.
W. Dong. Flocking of multiple mobile robots based on backstepping. IEEE Transactions on Systems, Man, and Cybernetics — Part B: Cybernetics, 2011, 41(2): 414–424.
J.-F. Hren, R. Munos. Optimistic planning of deterministic systems. Proceedings 8th European Workshop on Reinforcement Learning, Villeneuve d’Ascq, France: Springer, 2008: 151–164.
C. De Persis, P. Frasca. Robust self-triggered coordination with ternary controllers. IEEE Transactions on Automatic Control, 2013, 58(12): 3024–3038.
J. Mei, W. Ren, G. Ma. Distributed coordinated tracking with a dynamic leader for multiple Euler-Lagrange systems. IEEE Transactions on Automatic Control, 2011, 56(6): 1415–1421.
M. M. Zavlanos, G. J. Pappas. Distributed connectivity control of mobile networks. IEEE Transactions on Robotics, 2008, 24(6): 1416–1428.
M. Fiacchini, I. C. Morărescu. Convex conditions on decentralized control for graph topology preservation. IEEE Transactions on Automatic Control, 2014, 59(6): 1640–1645.
F. Bullo, J. Cortés, S. Martinez. Distributed Control of Robotic Networks: A Mathematical Approach to Motion Coordination Algorithms. Princeton: Princeton University Press, 2009.
J. Zhu, J. Lu, X. Yu. Flocking of multi-agent non-holonomic systems with proximity graphs. IEEE Transactions on Circuits and Systems, 2013, 60(1): 199–210.
H. Su, G. Chen, X. Wang, et al. Adaptive second-order consensus of networked mobile agents with nonlinear dynamics. Automatica, 2011, 47(2): 368–375.
J. Zhou, X. Wu, W. Yu, et al. Flocking of multi-agent dynamical systems based on pseudo-leader mechanism. Systems & Control Letters, 2012, 61(1): 195–202.
H. Tanner, A. Jadbabaie, G. Pappas. Flocking in teams of nonholonomic agents. Cooperative Control. V. Kumar, N. Leonard, A. Morse (eds.), Berlin: Springer, 2005: 458–460.
L. Buşoniu, C. Morărescu. Optimistic planning for consensus. Proceedings of the American Control Conference, Washington D.C.: IEEE, 2013: 6735–6740.
Lucian Buşoniu, C. Morărescu. Consensus for blackbox nonlinear agents using optimistic optimization. Automatica, 2014, 50(4): 1201–1208.
B. Jakubczyk, E. D. Sontag. Controllability of nonlinear discretetime systems: A Lie-algebraic approach. SIAM Journal of Control and Optimization, 1990, 28(1): 1–33.
R. S. Sutton, A. G. Barto. Reinforcement Learning: An Introduction. Cambridge, MA: MIT Press, 1998.
F. Lewis, D. Liu, eds. Reinforcement Learning and Approximate Dynamic Programming for Feedback Control. Hoboken: John Wiley & Sons, 2012.
R. Munos. The optimistic principle applied to games, optimization and planning: Towards foundations of Monte-Carlo tree search. Foundations and Trends in Machine Learning, 2014, 7(1): 1–130.
R. R. Negenborn, B. De Schutter, H. Hellendoorn. Multi-agent model predictive control for transportation networks: Serial versus parallel schemes. Engineering Applications of Artificial Intelligence, 2008, 21(3): 353–366.
J. Liu, X. Chen, D. M. de la Peña, et al. Sequential and iterative architectures for distributed model predictive control of nonlinear process systems. American Institute of Chemical Engineers (AIChE) Journal, 2010, 56(8): 2137–2149.
L. Buşoniu, R. Munos, Robert Babuška. A review of optimistic planning in Markov decision processes. Reinforcement Learning and Adaptive Dynamic Programming for Feedback Control. F. Lewis, D. Liu (eds.), Hoboken: John Wiley & Sons, 2012: DOI 10.1002/9781118453988.ch22.
L. Buşoniu, D. Ernst, B. De Schutter, et al. Approximate dynamic programming with a fuzzy parameterization. Automatica, 2010, 46(5): 804–814.
B. Kiumarsi, F. Lewis, H. Modares, et al. Reinforcement Qlearning for optimal tracking control of linear discrete-time systems with unknown dynamics. Automatica, 2014, 50(4): 1167–1175.
Author information
Authors and Affiliations
Corresponding author
Additional information
This work was supported by a Programme Hubert Curien-Brancusi cooperation grant (CNCS-UEFISCDI contract no. 781/2014 and Campus France grant no. 32610SE). Additionally, the work of L. Buşoniu was supported by the Romanian National Authority for Scientific Research, CNCS-UEFISCDI (No. PNII-RU-TE-2012-3-0040). The work of I.-C. Morărescu was partially funded by the National Research Agency (ANR) project “Computation Aware Control Systems” (No. ANR-13-BS03-004-02).
Lucian BUŞONIU received the M.Sc. degree (valedictorian) from the Technical University of Cluj-Napoca, Cluj-Napoca, Romania, in 2003, and the Ph.D. degree (cum laude) from the Delft University of Technology, Delft, the Netherlands, in 2009. He is an associate professor with the Department of Automation, Technical University of Cluj-Napoca, Romania. His research interests include planning-based methods for nonlinear optimal control, reinforcement learning and dynamic programming with function approximation, multi-agent systems, and, more generally, intelligent and learning techniques for control. He has authored a book as well as a number of journals, conferences, and chapter publications on these topics. Dr. Buşoniu was the recipient of the 2009 Andrew P. Sage Award for the best paper in the IEEE Transactions on Systems, Man, and Cybernetics.
Irinel-Constantin MORĂRESCU is currently an associate professor at Université de Lorraine and a researcher at the Research Centre of Automatic Control (CRAN UMR 7039 CNRS) in Nancy, France. He received the B.Sc. and the M.Sc. degrees in Mathematics from University of Bucharest, Romania, in 1997 and 1999, respectively. In 2006, he received the Ph.D. degree in Mathematics and Technology of Information and Systems from University of Bucharest and University of Technology of Compiègne, respectively. His works concern stability and control of time-delay systems, tracking control for nonsmooth mechanical systems, consensus and synchronization problems.
Rights and permissions
About this article
Cite this article
Buşoniu, L., Morărescu, IC. Topology-preserving flocking of nonlinear agents using optimistic planning. Control Theory Technol. 13, 70–81 (2015). https://doi.org/10.1007/s11768-015-4107-5
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11768-015-4107-5