Skip to main content

Secure Discrete-Time Linear-Quadratic Mean-Field Games

  • Conference paper
  • First Online:
Decision and Game Theory for Security (GameSec 2020)

Part of the book series: Lecture Notes in Computer Science ((LNSC,volume 12513))

Included in the following conference series:

Abstract

In this paper, we propose a framework for strategic interaction among a large population of agents. The agents are linear stochastic control systems having a communication channel between the sensor and the controller for each agent. The strategic interaction is modeled as a Secure Linear-Quadratic Mean-Field Game (SLQ-MFG), within a consensus framework, where the communication channel is noiseless, but, is susceptible to eavesdropping by adversaries. For the purposes of security, the sensor shares only a sketch of the states using a private key. The controller for each agent has the knowledge of the private key, and has fast access to the sketches of states from the sensor. We propose a secure communication mechanism between the sensor and controller, and a state reconstruction procedure using multi-rate sensor output sampling at the controller. We establish that the state reconstruction is noisy, and hence the Mean-Field Equilibrium (MFE) of the SLQ-MFG does not exist in the class of linear controllers. We introduce the notion of an approximate MFE (\(\epsilon \)-MFE) and prove that the MFE of the standard (non-secure) LQ-MFG is an \(\epsilon \)-MFE of the SLQ-MFG. Also, we show that \(\epsilon \rightarrow 0\) as the estimation error in state reconstruction approaches 0. Furthermore, we show that the MFE of LQ-MFG is also an \((\epsilon +\varepsilon )\)-Nash equilibrium for the finite population version of the SLQ-MFG; and \((\epsilon +\varepsilon ) \rightarrow 0\) as the estimation error approaches 0 and the number of agents \(n \rightarrow \infty \). We empirically investigate the performance sensitivity of the \((\epsilon +\varepsilon )\)-Nash equilibrium to perturbations in sampling rate, model parameters, and private keys.

Research support in part by Grant FA9550-19-1-0353 from AFOSR, and in part by US Army Research Laboratory (ARL) Cooperative Agreement W911NF-17-2-0196.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The red line in the boxplot represents the median, the box represents the 1st and 3rd quartile and the whiskers represent the max and min values, with outliers shown as red crosses.

References

  1. Moon, J., Başar, T.: Linear quadratic risk-sensitive and robust mean field games. IEEE Trans. Autom. Control 62(3), 1062–1077 (2016)

    Article  MathSciNet  Google Scholar 

  2. Breban, R., Vardavas, R., Blower, S.: Mean-field analysis of an inductive reasoning game: application to influenza vaccination. Phys. Rev. E 76(3), 031127 (2007)

    Article  Google Scholar 

  3. Couillet, R., Perlaza, S.M., Tembine, H., Debbah, M.: Electrical vehicles in the smart grid: a mean field game analysis. IEEE J. Sel. Areas Commun. 30(6), 1086–1096 (2012)

    Article  Google Scholar 

  4. Huang, M., Malhamé, R.P., Caines, P.E., et al.: Large population stochastic dynamic games: closed-loop Mckean-Vlasov systems and the Nash certainty equivalence principle. Commun. Inf. Syst. 6(3), 221–252 (2006)

    MathSciNet  MATH  Google Scholar 

  5. Lasry, J.-M., Lions, P.-L.: Mean field games. Japan. J. Math. 2(1), 229–260 (2007)

    Article  MathSciNet  Google Scholar 

  6. Huang, M., Caines, P.E., Malhamé, R.P.: Large-population cost-coupled LQG problems with nonuniform agents: individual-mass behavior and decentralized \(\varepsilon \)-Nash equilibria. IEEE Trans. Autom. Control 52(9), 1560–1571 (2007)

    Article  MathSciNet  Google Scholar 

  7. Bensoussan, A., Sung, K., Yam, S.C.P., Yung, S.-P.: Linear-quadratic mean field games. J. Optim. Theory Appl. 169(2), 496–529 (2016)

    Article  MathSciNet  Google Scholar 

  8. Huang, M., Zhou, M.: Linear quadratic mean field games-part I: the asymptotic solvability problem. arXiv preprint arXiv:1811.00522 (2018)

  9. Moon, J., Başar, T.: Discrete-time LQG mean field games with unreliable communication. In: 53rd IEEE Conference on Decision and Control, pp. 2697–2702. IEEE (2014)

    Google Scholar 

  10. Guo, X., Hu, A., Xu, R., Zhang, J.: Learning mean-field games. In: Advances in Neural Information Processing Systems (2019)

    Google Scholar 

  11. Fu, Z., Yang, Z., Chen, Y., Wang, Z.: Actor-critic provably finds Nash equilibria of linear-quadratic mean-field games. In: International Conference on Learning Representation (2020)

    Google Scholar 

  12. Elie, R., Pérolat, J., Laurière, M., Geist, M., Pietquin, O.: Approximate fictitious play for mean field games. arXiv preprint arXiv:1907.02633 (2019)

  13. Berberidis, D., Giannakis, G.B.: Data sketching for large-scale Kalman filtering. IEEE Trans. Signal Process. 65(14), 3688–3701 (2017)

    Article  MathSciNet  Google Scholar 

  14. Blocki, J., Blum, A., Datta, A., Sheffet, O.: The Johnson-Lindenstrauss transform itself preserves differential privacy. In: 2012 IEEE 53rd Annual Symposium on Foundations of Computer Science, pp. 410–419. IEEE (2012)

    Google Scholar 

  15. Tatikonda, S., Sahai, A., Mitter, S.: Stochastic linear control over a communication channel. IEEE Trans. Autom. Control 49(9), 1549–1561 (2004)

    Article  MathSciNet  Google Scholar 

  16. Kott, A., Swami, A., West, B.J.: The internet of battle things. Computer 49(12), 70–75 (2016)

    Article  Google Scholar 

  17. Janardhanan, S., Bandyopadhyay, B.: Output feedback sliding-mode control for uncertain systems using fast output sampling technique. IEEE Trans. Industr. Electron. 53(5), 1677–1682 (2006)

    Article  Google Scholar 

  18. Zaman, M., Zhang, K., Miehling, E., Başar, T.: Reinforcement learning in nonstationary discrete-time linear-quadratic mean-field games. In: 59th IEEE Conference on Decision and Control (2020, to appear)

    Google Scholar 

  19. Yang, Z., Chen, Y., Hong, M., Wang, Z.: Provably global convergence of actor-critic: a case for linear quadratic regulator with ergodic cost. In: Advances in Neural Information Processing Systems, pp. 8351–8363 (2019)

    Google Scholar 

  20. Zaman, M., Zhang, K., Miehling, E., Başar, T.: Approximate equilibrium computation for discrete-time linear-quadratic mean-field games. arXiv preprint arXiv:2003.13195 (2020)

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Muhammad Aneeq uz Zaman .

Editor information

Editors and Affiliations

7 Appendix

7 Appendix

In this section, we provide some necessary background material for completeness.

1.1 7.1 MFE of the LQ-MFG

Here we briefly discuss the MFE of the LQ-MFG which has been developed in a previous work [18]. We note from [18] that the dynamics of the generic agent in the LQ-MFG are given by,

$$\begin{aligned} x((k+1)\tau ) = A_0 x(k \tau ) + B_0 u (k \tau ) + w^0 (k \tau ) \end{aligned}$$
(24)

Although \(w^0\) is assumed to be non-Gaussian (Sect. 2.2) the results of [18] (which assume Gaussian distribution) still hold, since we restrict our attention to the class of linear controllers. In the standard LQ-MFG, the multi-rate setup is not required since the controller has access to the true state of the agent. The generic agent aims to minimize the cost function,

$$\begin{aligned} \tilde{J}(\mu ,\bar{x}) = \limsup _{T \rightarrow \infty } \frac{1}{T} \mathbb {E}_{\mu } \Big \{ \sum _{k=0}^{T-1} ||x(k \tau ) - \bar{x}(k \tau ) ||_{Q}^2 + ||u(k \tau )||_R^2 \Big \}, \end{aligned}$$
(25)

where \(\bar{x}\) is the mean-field trajectory. Next we restate the existence and uniqueness guarantees of MFE for the LQ-MFG.

Proposition 1

([18]). Under Assumption 1 the LQ-MFG ((24)–(25)) admits the unique MFE given by the tuple \((K^*,F^*) \in \mathbb {R}^{p \times 2m} \times \mathbb {R}^{m \times m}\). The matrix \(F^* = \varLambda (K^*) = A_0 - B_0(K^*_1 + K^*_2)\), and controller \(K^*\) is defined as,

$$\begin{aligned} K^* = (\bar{B}^T P^* \bar{B} + R )^{-1} \bar{B}^T P^* \bar{A}^*, \text { where } \bar{A}^* = \begin{bmatrix} A_0 &{} 0 \\ 0 &{} F^* \end{bmatrix} \end{aligned}$$
(26)

and \(P^*\) is the solution to the DARE,

$$\begin{aligned} P^* = \bar{A^*}^T P^* \bar{A^*} + \bar{Q} - \bar{A^*}^T P^* \bar{B}(R + \bar{B}^T P^* \bar{B})^{-1} \bar{B}^T P^* \bar{A^*} \end{aligned}$$
(27)

and \(\bar{B}\) and \(\bar{Q}\) as defined in (11) and (12), respectively.

The DARE is obtained by substituting \(K^*\) in the Lyapunov Eq. (16) hence \(P^* = P_{K^*}\). An important point to note is that in the LQ-MFG the estimation error is 0, as the controller has perfect access to the state of the agent. This translates to the covariance matrix of estimation error \(\varSigma _C = 0\) and hence \(\hat{\varSigma }_C = 0\) for LQ-MFG. Using (15) the cost of linear controller K and linear trajectory defined by matrix F for the LQ-MFG will be

$$\begin{aligned} \tilde{J}(K,F) = \mathrm{Tr}\,(P_K \bar{\varSigma }) \end{aligned}$$
(28)

where \(P_K\) is the solution to the Lyapunov Eq. (15). Furthermore it can also be verified that

$$\begin{aligned} K^* = \mathrm{argmin}_K \tilde{J} (K, F^*) \end{aligned}$$
(29)

This MFE \((K^*,F^*)\) can be obtained by using the mean-field update operator as discussed in [18].

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

uz Zaman, M.A., Bhatt, S., Başar, T. (2020). Secure Discrete-Time Linear-Quadratic Mean-Field Games. In: Zhu, Q., Baras, J.S., Poovendran, R., Chen, J. (eds) Decision and Game Theory for Security. GameSec 2020. Lecture Notes in Computer Science(), vol 12513. Springer, Cham. https://doi.org/10.1007/978-3-030-64793-3_11

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-64793-3_11

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-64792-6

  • Online ISBN: 978-3-030-64793-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics