# Renormalization of Collective Modes in Large-Scale Neural Dynamics

## Abstract

The bulk of studies of coupled oscillators use, as is appropriate in Physics, a global coupling constant controlling all individual interactions. However, because as the coupling is increased, the number of relevant degrees of freedom also increases, this setting conflates the strength of the coupling with the effective dimensionality of the resulting dynamics. We propose a coupling more appropriate to *neural circuitry*, where synaptic strengths are under biological, activity-dependent control and where the coupling strength and the dimensionality can be controlled separately. Here we study a set of \(N\rightarrow \infty \) strongly- and nonsymmetrically-coupled, dissipative, powered, rotational dynamical systems, and derive the equations of motion of the reduced system for dimensions 2 and 4. Our setting highlights the statistical structure of the eigenvectors of the connectivity matrix as the fundamental determinant of collective behavior, inheriting from this structure symmetries and singularities absent from the original microscopic dynamics.

## Keywords

Renormalization Neural networks Coarse-graining Neural dynamics Cortical models## 1 Introduction

Equilibrium statistical mechanics is static in nature, its program being the reduction of a thermodynamically-large system to a handful of relevant variables and their equations of state. A major subprogram of *nonequilibrium* statistical mechanics consists of reducing a thermodynamically-large system to a handful of relevant variables *and their dynamical laws*.

We can describe this schematically as understanding under which conditions the the diagram in Fig. 1 commutes, i.e. when taking either branch leads to the same result.

At the top left we have our thermodynamically-large system, which lives in dimension \(N_H\), ruled by a high-dimensional dynamical law. At the bottom right we have a low-dimensional space, the reduced variables, which live in dimension \(N_L\). On the left branch, we first perform a detailed, costly and precise simulation of every variable in the system, and *then* we project this high-dimensional solution down into the lower-dimensional space. On the right branch, instead, we *first* project down the equations of motion themselves, to obtain an effective dynamics on the lower-dimensional subspace. Then we evolve the effective dynamics in its low-dimensional space. If we get the same low-D trajectories (or equivalent trajectories in some appropriately-defined sense) as a result of both branches, then we can say we have reduced our system.

We shall carry out this program for a particular setting within the theory of neural networks. A number of studies have highlighted the recurrent intracortical connections as a major determinant of the patterns of global activity within the cerebral cortex. For example, the spontaneous activity of individual units in primary visual cortex is strongly coupled to global patterns evincing the underlying functional architecture [12]: during spontaneous activity in the absence of input, global modes are transiently activated that are extremely similar to modes engaged when inputs are presented. Similarly in motor control it has been shown both in motor reach actions [13] as well as in songbirds [14] that motor gestures result in the transient activation of extended patterns of coherent activity. We have previoiusly argued that homeostatic mechanisms, such as those seeking to balance excitation and inhibition, will result in many dynamical modes of the system posing themselves close to the boundary between stability and instability [15]. Near the onset of instability, the spatial layout of such dynamical modes is determined as eigenvectors of a connectivity matrix [16]; their transient activation is then consistent with the eigenvalue of such an eigenvector being driven through the instability line, and back [14, 17].

We will try to therefore take a point of view in which we have as our two ingredients a cortical circuit, and a pattern of intracortical connections between copies of the same circuit. We thus consider an extremely abstract idealization of the connectivity of the cerebral cortex, in which we have cortical micro-columns (thought of, generally, as the vertical direction) with a prescribed complex intra-columnar circuit, laid out on a surface (thought of, generally, as horizontal). In general we shall have the neurons distributed on the product of a space describing the cortical column (say the interval *I*) times a space describing the surface (say the two-d plane \(\mathbb R^{2}\)). In general we shall have connections abstracted so that there are only connections within a column, or between homologous elements across columns, within the same layer; i.e. only horizontal or vertical connections, never diagonal. (This is reminiscent of some high-throughput computing architectures used for massively-parallel computing, such as the Anton supercomputer [18]). In keeping with a long tradition in computational neuroscience, we shall model the intercolumnar connectivity as being random [19]. Therefore our model will be on one side highly structured (the given cortical circuit) and on the other side entirely structureless (the random connectivity across columns).

The structure of this paper is as follows. We shall first introduce definitions and general framework. Then we introduce a specific simple example that under some circumstances we can actually solve in pretty closed form: a set of coupled oscillators. We will address increasingly complicated subproblems based on the nature of the coupling matrix: first, we will control a single real eigenvalue of a normal matrix, will give us a global equation in two degrees of freedom. Then we will study two complex-conjugate eigenvalues; this will give us a global equation in 4 degrees of freedom. Then we will show that there are interesting issues associated to having a non-normal connectivity matrix, namely one whose eigenvectors are not orthogonal. Finally, we will examine the explicit form of the mapping between microscopic and macroscopic modes, and show how it changes the smoothness class of the dynamics, making the global mode dynamical system less regular than the microscopic one. A careful study of this change in regularity permits us to give an explicit, closed-form answer as to under which conditions in the microscopic dynamics the program outlined in Fig. 1 can be successfully completed. At the end, we shall have given an explicit set of microscopic dynamics for which a proper global dynamics exists, and we shall compute this global dynamics explicitly.

## 2 Historical Note

This paper is dedicated to the memory of Leo P. Kadanoff, who was the postdoctoral mentor of OP, and the PhD thesis advisor of MM. It would be hard to express how much Leo shaped, not just this field, but the very process of how the authors reason, how we choose worthy problems, and how we try to train the younger generation.

In particular, Leo always made emphasis on using simple, highly stylized models as caricatures of a more general natural phenomenon one tried to understand, always keeping in mind *what is the underlying program* one is trying to push forward. His use of spin blocking to bring to the forefront the notion of averaging only over fluctuations at a given scale is emblematic, jokingly dubbed by his close friend Ken Wilson as an idea that “*...although absurd, will be the basis of generalizations which are not absurd*” [20]. The elegance with which such a simple setting shows that universal quantities like critical exponents are not a property of the underlying system, but rather a property (in this case an eigenvalue) of a change-of-scale operator, is a trademark of his work in general. Following Leo’s teachings, we will in this Paper outline a general program, and then solve very explicitly one small example in the program as to illustrate exactly where the conceptual emphasis should go.

We’ll miss his contributions as colleague and teacher. We shall also miss his personality and even his mannerisms. Leo was an astoundingly quick thinker and had a deep understanding, not just of Physics, but also of how physicists go about trying to solve problems. So during scientific discussions he often could foresee well ahead what was coming, often leaving his disciples with the feeling of having just been X-rayed. In the rare occasions one managed to catch him by surprise at a key point in the argument, he would lean back, briefly stare at the celing and proffer a curt yet solid “Ha!”, which was a much-sought and vied-for high mark of approval among us. It deeply saddens us we may never again receive that particular honor.

## 3 Definitions

Let us now establish the notation to be used in further detail, shown schematically in Fig. 3.

Again, at the top left we have our thermodynamically-large system: many variables \(\mathbf{x} \in {\mathbb R}^{N_H}\) ruled by a high-dimensional dynamical law \(\dot{\mathbf{x}} = f(\mathbf{x})\) given by a vectorfield *f* in \( {\mathbb R}^{N_H}\), stated as \(f \in {\mathcal T}({\mathbb R}^{N_H})\). At the bottom right we have a low-dimensional space, the reduced variables, which live in \({\mathbb R}^{N_L}\). On the left branch we first evolve in \(N_H\) and *then* we use a projection operator \(\Pi \) to project the high-dimensional solution down into the lower-dimensional space. On the right branch, instead, we *first* project down the equations of motion, to obtain an effective dynamics *F* on the lower-dimensional subspace: \(F \in {\mathcal T}({\mathbb R}^{N_L})\). We call the projector operator for the equations \(\Phi \); then \(F = \Phi \{f\}\). Then we evolve this simpler dynamics \(\dot{X} = F(X)\). If we get the same trajectories (or equivalent trajectories in some well-defined sense) as a result of both branches, then we can say we have reduced our system.

*x*,

*y*and

*z*, and our differential equation connects them vertically to one another through a 3D vector field (

*f*,

*g*,

*h*):

*N*columns, we get

*N*identical sets of uncoupled differential equations, one per column:

*x*with a linear synaptic matrix

*M*

*N*linearly coupled identical 3

*D*equations in 3

*N*variables. Much has been written about such systems, particularly in the area of coupled oscillators.

A common way to study such systems is to define \(M_{ij} = \lambda G_{ij}\), where \(\lambda \) is a global coupling constant and \(G_{ij}\) a graph connectivity matrix, whose elements are either 0 or 1, which can be used to describe the connections as occurring only along an underlying lattice. This is a standard setting in Physics, where the interaction strength is controlled by a physical constant and therefore affects all extant interactions in parallel. This approach leads to a changing effective dimensionality of the dynamical system, because as \(\lambda \) is varied, all dynamical eigenvalues of the system move in unison, and hence the number of modes which can potentially enter the dynamics increases with increasing \(\lambda \). We will take a different approach.

We shall assume the matrix *M* is under control from slow homeostatic processes, and that the stability of different eigenmodes is under biological control. To approach this goal in a graded fashion we shall begin by assuming that most of the eigensystem \((\lambda ^i,\mathbf{e}^i) \ni M \mathbf{e} = \lambda \mathbf{e} \) has strongly negative values. In addition we shall place the more specific hypothesis on the vector field (*f*, *g*, *h*) that when all eigenvalues \(\lambda \) of *M* are sufficiently negative there exists a single, stable fixed point of the dynamics.

Then we shall allow either one single real eigenvalue or a couple of complex-conjugate eigenvalues to increase and approach the stability limits from the left.

*N*(spectrum in the unit disk in the complex plane, matrix almost surely non-normal) or the antisymmetric component of said \(M_0\), \(M_a = (M_0 - M^t_0)/\sqrt{2}\) (purely imaginary spectrum, matrix is normal). Then, a specific eigenvalue \(\lambda _1\) (or pair of complex-conjugate eigenvalues) associated to an eigenvector \(\mathbf e\) (again, pair of), is singled out and

*N*columns and

*L*layers, then its size is

*NL*. The direction across layers is the rigidly-defined cortical circuit, and linear stability analysis may be performed on this circuit to obtain a Hessian matrix whose

*L*eigenvalues determine the stability of one single column in isolation. The direction across columns is defined by an entirely unstructured random matrix, having

*N*eigenvalue/eigenvector pairs defining the modes of behavior of the system. The system as a whole, then, has a stability structure given by

*NL*eigenvalues that is a complex interplay of the structured column and the unstructured connections. For concreteness, assuming \(L=3\), the system as a whole is given by a vectorfield of dimension 3

*N*. The stability matrix of the whole system, namely the derivative matrix of the vectorfield, is then given by a matrix

*J*of size \(3N \times 3N\), which we can group as \(3\times 3\) blocks of size \(N \times N\), eight of which are diagonal:

*J*from those of

*M*will be discussed elsewhere.

## 4 A Concrete Example: 2D Dynamics

*f*(

*x*).

*f*order-by-order, so some of what follows concentrates on \(f(x)=x^\alpha \). But the main issue is the structure of

*M*, whether it’s normal or non-normal; and whether we control a single real or two complex-conjugate eigenvalues.

As we shall see shortly, our procedure for obtaining the global dynamics from the microscopic dynamics involves both the right and the left eigenvectors of *M*. In general the matrix of left eigenvectors is the inverse of the matrix of right eigenvectors. However if *M* is normal, then the inverse reduces to the complex conjugate making their relationship much simpler. Therefore we shall explore first normal matrices and then examine the additional issues raised by non-normality. Within normal matrices we shall, as anticipated in Fig. 4, either control a single real eigenvalue or two complex-conjugate ones.

### 4.1 Single Real Eigenvalue, Normal Matrix

The simplest sub-case of Eq. (4) is that of a normal matrix *M* and a single real eigenvalue under control.

*M*to be given by Eq. 2 with \(M_a\), the antisymmetric component of a matrix of \(N\times N\) i.i.d. Gausian variables of std \(1/\sqrt{N}\); its eigenvalues are purely imaginary \(\lambda = i \omega \) with imaginary component distributed like \(p(\omega )\approx \sqrt{1-\omega ^2/4}\). Consider one of its null eigenvalues (one is guaranteed to exist if

*N*is odd) and the corresponding eigenvector \(\mathbf e \). Then define as per Eq. (2)

*M*, except for the null one we chose, get moved to \(-\varLambda \), while the null one now has value \(\lambda \).

*X*is of order 1, because \(\mathbf{x \cdot x} \approx N\) while \(\mathbf e \cdot e = 1\) so that \( \mathbf{x \cdot e } \approx \sqrt{N}\) generically.

*f*, meaning we can try to apply it order-by-order in a Taylor expansion of

*f*. Applying Eq. (9) to individual integer powers yields, in the large

*N*limit, an evaluation of the moments of a Gaussian distribution:

*independent*of which eigenvector \(\mathbf{e}\) we chose, because under these circumstances the operator is self-averaging. Then, the power-law is unchanged except for a prefactor: the cubic \(X^3\) nonlinearity gets renormalized by a factor of 3, \(X^5\) by 15, \(X^7\) by 105. Curiously this operator

*destroys all even nonlinearities*. In subsequent work we shall prove that the even nonlinearities get absorbed in the overall shape of the center manifold, which we have assumed in the ansatz Eq. (7) to be linear in \(x_i\). However the reduced equations (8) are not affected.

### 4.2 Complex-Conjugate Pair of Eigenvalues, Normal Matrix

As shown in Fig. 4, the second-simplest case is to control two complex-conjugate eigenvalues, as they can be controlled by a single parameter, their real part, while leaving their imaginary part unchanged.

We follow the previous section in using \(M_a\), a normal matrix, as our base connectivity structure. Now we choose a pair of complex-conjugate eigenvalues. The corresponding eigenvectors are also complex conjugate. Choosing as \(\mathbf{e}\) the one corresponding to the eigenvalue with positive imaginary part, the 2D subspace is spanned both by \(\mathbf{e}\) and \( \bar{\mathbf{e}}\), or by \(\mathbf{e}_R \equiv \mathfrak {R}\mathbf{e}\) and \(\mathbf{e}_I \equiv \mathfrak {I}\mathbf{e}\).

*X*, now complex, as the normalized projection on \(\mathbf{e}\):

*X*equation is now a Hopf bifurcation normal form in a single complex variable, where \(\mathrm{Imag}(\lambda )\) is the rotational frequency in the

*X*complex plane, and \(\mathrm{Real}(\lambda )\) is the control parameter, while the whole

*XY*plane equation uses the same \(\mathrm{Real}(\lambda )\) as a control parameter for a Hopf bifurcation in the real

*XY*plane. Thus this equation contains 2 distinct frequencies, \(\omega = 1\) from the base equation, and \(\omega = \mathrm{Imag}(\lambda )\), the frequency of the eigenvector, from the complex

*X*equation, and both cross the threshold of stability at the same time. Therefore, in this collective mode, a codimension-2 transition becomes generic, and the system in principle can bifurcate directly from a fixed point onto a torus. In this sense it is similar to a Takens–Bogdanoff normal form, only that it has two distinct and independent frequencies.

Thus the macroscopic equations are non-generic with respect to perturbations in the underlying microscopic dynamics in two ways: first, all even terms are destroyed, and second it can undergo a direct transition from a fixed-point to a torus.

### 4.3 Non-normal Matrices

In general, matrices are non-normal. Therefore to study general matrices we need to learn how to understand how non-normality affects our strategy.

For normal matrices, eigenvectors are orthogonal and hence the matrix of left eigenvectors, the inverse of the matrix of right eigenvectors, can be simply obtained as the adjunct (complex conjugate transposed) of the right. However, when the matrix is not normal, the matrix of left eigenvectors has to be obtained the hard way, by inverting the matrix of right eigenvectors; the most important thing that happens is that the left eigenvectors are no longer normalized to \(|e^\dag |=1\).

We now study what happens when, instead of choosing our base connectivity matrix to be the antisymmetric \(M_a\), which is normal, we choose the standard random matrix \(M_0\), which is non-normal. The eigenvalue expansion of \(M_0\) is \(E_R D E_L\) where *D* is a diagonal matrix with the eigenvalues on the diagonal, \(E_R\) is the matrix of the right eigenvectors, and \(E_L\), the matrix of left eigenvectors, satisfies \(E_R E_L = I\).

Quantitatively this stems from a simple reason. The right eigenvectors of \(M_a\), by normality, must be a unitary matrix and thus all its eigenvalues lie on the unit circle in the complex plane. However, the right eigenvectors of \(M_0\) are not orthogonal to one another, but they are still normalized to unity, reducing *drastically* the determinant of \(E_R\). This reduction can only be due to the eigenvalues of \(E_R\) lying inside the unit circle.

*eigenvectors*of said matrix go to zero much faster, with the log-linear plot displaying significant convexity. Furthermore, the probability distribution of the log said determinant is rather broad, as shown in Fig. 7.

## 5 Singularities

We have seen that our \(\Phi \) operator can be applied to the vectorfield *f* order-by-order in a Taylor expansion. We now need to understand whether the resulting series converges and whether it can be resummed, in order to know whether we have succeeded in coarse-graining the dynamics. As we shall see, there exist *f* for which this can be done, and there exist *f* for which it can’t. In order to establish the boundary between them we need to understand how \(\Phi \) affects the singularity structure of *f*.

It is a textbook point that in statistical mechanics the global behavior in the thermodynamic limit acquires singularities not present at finite *N*, which are phase transitions and other exotic objects. More technically, the partition function is assured smooth for finite *N* for local bounded interactions, but under some conditions acquires singular behavior for infinite *N*. As an exception, van Hove’s theorem assures us local interactions in a one-dimensional system cannot give rise to a phase transition at finite nonzero temperature.

In keeping with this general idea, in the soluble cases presented above, the \(\Psi \) operator increases the singularity class of the vector field, by multiplying every odd term in the Taylor expansion with the double-factorial of the power.

Since the double factorial scales approximately like the square root of the factorial itself, this has peculiar consequences. If the function *f* has singularities, then \(\Psi \) magnifies them so much that it may fail to converge at all. If the terms of the expansion of *f* decrease like 1 / *n*! (i.e. an exponential or sine), then the terms of *F* decrease like 1 / *n*!! (such as \(\sin x^2\)) and *F* is smooth.

*f*decrease like 1 /

*n*!!, in which case \(\Psi \) generates a simple singularity. As an example,

Thus the peculiar structure of \(\Phi \) permits us to make general statements about under which vectorfields will the diagram of Fig. 3 commute.

## 6 Context

When pitted against the seemingly unfathomable depth of complexity within living beings, it is a natural “zero information” approach to modeling to assume the structure of interactions is, for all practical purposes of the modeler, *structureless*, a jumble of random numbers and noise. A classical embodiment of this line of thought are Kauffman nets [21, 22], a modelization of transcriptional regulation networks as random boolean nets that was instrumental in highlighting the intrinsically dynamical nature of cell differentiation. In this family of models, the complex web of transcriptional regulation is described as a random boolean network, its fixed-point attractors representing steady, self-consistent states of transcriptional activation that Kauffman equates to stable cell fate types; differentiation then being a process of climbing over metastable barriers separating basins of adjacent fixed points.

In neuroscience, a fruitful line of inquiry started with the Sompolinsky family of neural net models, which assume as a basis an interaction structure given by a Gaussian random matrix scaled to have a spectrum in the unit disk, and the overall excitatory/inhibitory balance set so the unit disk is displaced to the left until it osculates the imaginary axis [19]. These results have been particularly instrumental as a beachhead in the study of *recurrent* neural networks, which have otherwise proved to be resilient to analysis.

Using this connectivity as their basis, a long series of works shows that by careful tampering with such a structure one may achieve a number of useful ends, with a notable flurry of recent activity [23, 24, 25, 26, 27, 28, 29, 30, 31].

In previous work [15] it was shown that certain slow dynamics to evolve the connectivity structure itself, such as anti-Hebbian evolution, could result in drastic re-shaping of the eigenvalue distribution, where eigenvalues were flattened onto the imaginary axis, the edge of stability for the model.

## 7 Conclusions

There are a number of points to be made at this juncture. First, that under certain circumstances this procedure converges properly and the diagram of Fig. 1 commutes: we can obtain in closed form the coarse-grained dynamics.

Second, our reduction procedure depends crucially on the structure of the *eigenvectors*. After more than 60 years of interest in eigenvalue statistical structure, and a tour-de-force theorem [32] for the circular law, a lot is now understood about them. Still remarkably little is known about the structure of the eigenvectors of such matrices, which display much less universality than the eigenvalues; [33] was one of the first studies of eigenvectors of nonnormal random matrices. Basic results link the PDF of entries in the eigenvector to the PDF of the matrix entries [34], and a fair amount of effort has been spent in understanding localization of eigenvectors in nearly-one-dimensional systems; see [35] for a recent review.

Simple numerical exploration leads us to conjecture that a great deal about *the geometrical structure of the network gets encoded into the statistical structure of the eigenvalues*. To illustrate why, let us consider a simple example: a 2D toroidal (periodic boundary conditions) lattice of size \(M\times M\), in which elements are connected by a random Gaussian term if they are \({\le }3\) lattice units apart. There are 28 such near neighbors for each site. There are \(N=M^2\) sites, and hence the interaction matrix is \(N\times N = M^4\); this matrix is *sparse* because every row contains 29 nonzero elements (including the diagonal) no matter how large *M* is. Since the banded structure of the matrix exceeds \(\sqrt{N}=M\) this construction narrowly escapes localization [35]. The spectrum of this matrix is still circular, following the strong universality results for the spectrum of sparse matrices; but the eigenvectors have a strongly interdependent statistical structure. The reason is simple: we are using \(29 M^2\) random numbers are used in order to generate \(M^2\) eigenvalues and \(M^2\) eigenvalues with \(M^2\) coefficients (minus one for normalization). Therefore the eigenvectors are not merely correlated: they are *not statistically independent from one another*, and they become more and more codependent the larger the value of *M*. Our reduction operator is extremely sensitive to these dependencies, and we conjecture that the ambient space in which the columns are embedded gets encoded into these interdependencies.

Third, this procedure generates symmetries where there were no symmetries to begin with. In the current case, the global dynamics is always an odd vector field. This can lead to curious situations, in which the dynamics of the individual column is unstable (say due to a \(x^2\) or \(x^4\) term), but the reduced dynamics is a globally attractive limit cycle.

Finally, in classical statistical mechanics a notorious feature is that the system as a whole can develop singularities such as phase transitions where the underlying system was perfectly smooth. Like in statistical mechanics, the global dynamics is of a slightly smaller regularity class than the microscopic dynamics, as multiplication by the double-factorial of *N* makes the Taylor expansion less convergent. For example, if the microscopic dynamics *f* has a pole, then *F* is nowhere-convergent and the dynamics is not coarse-grainable through the method presented in this Paper. If, on the other hand, *f* is holomorphic, it may stay holomorphic because the double factorial is approximately the square root of the factorial for sufficiently large arguments. On the other hand certain holomorphic functions get mapped onto functions with poles, increasing their singularity class.

We shall present elsewhere a more rigorous approach using center manifold reductions to a higher order than here, and show that it accurately captures numerical simulations. In further work we shall concentrate on controlling multiple modes and deriving their laws of interaction, as well as examining matrices *M* with more interesting structure, such as locality in an embedding ambient space.

## Notes

### Acknowledgements

OP acknowledges Grants FIS2013-48444-C2-1-P/MINECO and FIS2016-77692-C2-1-P/MINECO/AEI/FEDER,UE as well as the Program of Mobility of MINECO.

## References

- 1.Kadanoff, L.P., Swift, J.: Transport coefficients near the critical point: a master-equation approach. Phys. Rev.
**165**(1), 310–322 (1968)ADSCrossRefGoogle Scholar - 2.Ott, E., Antonsen, T.M.: Low dimensional behavior of large systems of globally coupled oscillators. Chaos
**18**(3), 037113 (2008). arXiv:0806.0004 ADSMathSciNetCrossRefMATHGoogle Scholar - 3.Boldrighini, C., Franceschini, V.: 5-Dimensional truncation of the plane incompressible Navier–Stokes equations. Commun. Math. Phys.
**64**(2), 159–170 (1979)ADSMathSciNetCrossRefMATHGoogle Scholar - 4.Franceschini, V., Tebaldi, C.: Sequences of infinite bifurcations and turbulence in a five-mode truncation of the Navier–Stokes equations. J. Stat. Phys.
**21**(6), 707–726 (1979)ADSMathSciNetCrossRefGoogle Scholar - 5.Franceschini, V., Tebaldi, C.: A seven-mode truncation of the plane incompressible Navier–Stokes equations. J. Stat. Phys.
**25**(3), 397–417 (1981)ADSMathSciNetCrossRefGoogle Scholar - 6.Landau, L.D., Lifshitz, E.M.: Fluid Mechanics, Course of Theoretical Physics, vol. 6. Pergamon Press, Oxford (1987)Google Scholar
- 7.Landau, L.D., Lifshitz, E.M.: Statistical Physics, Course of Theoretical Physics, Vol. 5, Chap. XIV. Pergamon Press, Oxford (1994)Google Scholar
- 8.Ginzburg, V.L.: Some remarks on phase transitions of the 2nd kind and the microscopic theory of ferroelectric materials. Sov Phys Solid State
**2**(9), 1824–1834 (1961)Google Scholar - 9.Cross, M., Greenside, H.: Pattern Formation and Dynamics in Nonequilibrium Systems. Cambridge University Press, Cambridge (2009)CrossRefMATHGoogle Scholar
- 10.Cross, M., Hohenberg, P.C.: Pattern formation outside of equilibrium. Rev. Mod. Phys.
**65**(3), 851–1112 (1993)ADSCrossRefGoogle Scholar - 11.Walgraef, D.: Spatio-Temporal Pattern Formation, with Examples in Physics, Chemistry and Materials Science. Springer, New York (1996)MATHGoogle Scholar
- 12.Tsodyks, M., Kenet, T., Grinvald, A., Arieli, A.: Linking spontaneous activity of single cortical neurons and the underlying functional architecture. Science
**286**(5446), 851–1112 (1999). doi: 10.1126/science.286.5446.1943 CrossRefGoogle Scholar - 13.Churchland, M.M., Cunningham, J.P., Kaufman, M.T., Foster, J.D., Nuyujukian, P., Ryu, S.I., Shenoy, K.V.: Neural population dynamics during reaching. Nature
**487**(7405), 51–56 (2012)ADSGoogle Scholar - 14.Amador, A., Sanz-Peri, Y., Mindling, G.B., Margoliash, D.: Elemental gesture dynamics are encoded by song premotor cortical neurons. Nature
**495**, 59–64 (2013). doi: 10.1038/nature11967 ADSCrossRefGoogle Scholar - 15.Magnasco, M.O., Piro, O., Cecchi, G.A.: Self-tuned critical anti-Hebbian networks. Phys. Rev. Lett.
**102**(25), 258102 (2009)ADSCrossRefGoogle Scholar - 16.Yan, X.H., Magnasco, M.O.: Input-dependent wave attenuation in a critically-balanced model of cortex. PLoS ONE
**7**(7), e41419 (2012)ADSCrossRefGoogle Scholar - 17.Gardner, T., Cecchi, G., Magnasco, M., Laje, R., Mindlin, G.B.: Simple motor gestures for birdsongs. Phys. Rev. Lett.
**87**(20), 208101 (2001). doi: 10.1103/PhysRevLett.87.208101 ADSCrossRefGoogle Scholar - 18.Shaw, D.E.: Architectures and algorithms for millisecond-scale molecular dynamics simulations of proteins. Proceedings of the 41st IEEE/ACM International Symposium on Microarchitecture (2008). doi: 10.1109/MICRO.2008.4771773
- 19.Sompolinsky, H., Crisanti, A., Sommers, H.J.: Chaos in random neural networks. Phys. Rev. Lett.
**61**(3), 259–262 (1988)ADSMathSciNetCrossRefGoogle Scholar - 20.Wilson, K.G.: Renormalization group and critical phenomena. I. Renormalization group and the kadanoff scaling picture. Phys. Rev. B.
**4**(9), 3174–3183 (1971)ADSCrossRefMATHGoogle Scholar - 21.Kauffman, S.: Metabolic stability and epigenesis in randomly constructed genetic nets. J. Theor. Biol.
**22**(3), 437–467 (1969). doi: 10.1016/0022-5193(69)90015-0. PMID 5803332 - 22.Kauffman, S.A., McCulloch, W.S.: Random nets of formal genes. Quarterly Progress Report 34. Research Laboratory of Electronics, Massachusetts Institute of Technology (1967)Google Scholar
- 23.Stern, M., Sompolinsky, H., Abbott, L.F.: Dynamics of random neural networks with bistable units. Phys. Rev. E
**90**, 062710 (2014)ADSCrossRefGoogle Scholar - 24.Lalazar, H., Abbott, L.F., Vaadia, E.: Tuning curves for arm posture control in motor cortex are consistent with random connectivity. PLoS Comput Biol
**12**(5), e1004910 (2016)ADSCrossRefGoogle Scholar - 25.Rajan, K., Harvey, C.D., Tank, D.W.: Recurrent network models of sequence generation and memory. Neuron
**90**, 1–15 (2016)CrossRefGoogle Scholar - 26.Ostojic, S.: Two types of asynchronous activity in networks of excitatory and inhibitory spiking neurons. Nat. Neurosci.
**17**, 594–600 (2014)CrossRefGoogle Scholar - 27.Kadmon, J., Sompolinsky, H.: Transition to chaos in random neuronal networks. Phys. Rev. X
**5**, 041030 (2015)Google Scholar - 28.Landau, I., Egger, R., Dercksen, V.J., Oberlaender, M., Sompolinsky, H.: The impact of structural heterogeneity on excitation-inhibition balance in cortical networks. Neuron
**92**(5), 1106–1121 (2016)CrossRefGoogle Scholar - 29.Sompolinsky, H.: Computational neuroscience: beyond the local circuit. Curr. Opin. Neurobiol.
**25**, 1–6 (2014)CrossRefGoogle Scholar - 30.Engelken, R., Farkhooi, F., Hansel, D., Vreeswijk, C., Wolf, F.: A reanalysis of “Two types of asynchronous activity in networks of excitatory and inhibitory spiking neurons”. F1000Research
**5**, 2043 (2016)CrossRefGoogle Scholar - 31.Harish, O., Hansel, D.: Asynchronous rate chaos in spiking neuronal circuits. PLoS Comput. Biol.
**11**(7), e1004266 (2015)ADSCrossRefGoogle Scholar - 32.Tao, T., Vu, V.: Random matrices: the circular Law. Commun. Contemp. Math.
**10**(2), 261–307 (2008). arXiv:0708.2895 - 33.Chalker, J.T., Mehlig, B.: Eigenvector statistics in non-Hermitian random matrix ensembles. Phys. Rev. Lett.
**81**, 3367 (1988)ADSCrossRefGoogle Scholar - 34.Tao, T., Vu., V.: Random matrices: Universal properties of eigenvectors. Random Matrices Theory Appl
**1**(01), 1150001 (2012). arXiv:1103.2801v2 - 35.O’Rourke, S., Vu, V., Wang, K.: Eigenvectors of random matrices: a survey. arXiv:1601.03678 (2016)

## Copyright information

**Open Access**This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.