Abstract
Dynamics and function of neuronal networks are determined by their synaptic connectivity. Current experimental methods to analyze synaptic network structure on the cellular level, however, cover only small fractions of functional neuronal circuits, typically without a simultaneous record of neuronal spiking activity. Here we present a method for the reconstruction of large recurrent neuronal networks from thousands of parallel spike train recordings. We employ maximum likelihood estimation of a generalized linear model of the spiking activity in continuous time. For this model the point process likelihood is concave, such that a global optimum of the parameters can be obtained by gradient ascent. Previous methods, including those of the same class, did not allow recurrent networks of that order of magnitude to be reconstructed due to prohibitive computational cost and numerical instabilities. We describe a minimal model that is optimized for large networks and an efficient scheme for its parallelized numerical optimization on generic computing clusters. For a simulated balanced random network of 1000 neurons, synaptic connectivity is recovered with a misclassification error rate of less than 1 % under ideal conditions. We show that the error rate remains low in a series of example cases under progressively less ideal conditions. Finally, we successfully reconstruct the connectivity of a hidden synfire chain that is embedded in a random network, which requires clustering of the network connectivity to reveal the synfire groups. Our results demonstrate how synaptic connectivity could potentially be inferred from largescale parallel spike train recordings.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
The synaptic organization of neuronal networks is key to understanding the dynamics of brain circuits, and, eventually, to link them to higher level cognitive functions. A large body of work aims to address this challenge by developing experimental techniques which enable the reconstruction of the connections between neurons on the basis of anatomical or physiological evidence. Anatomically, synaptic connections may be identified using optical imaging or electron microscopy (Briggman et al. 2011; Bock et al. 2011), while physiological approaches rely on simultaneous recordings of individual neurons and the mutual influence of the spikes of one neuron on the membrane potential of the other (Perin et al. 2011; Boucsein et al. 2011). Substantial progress has been made in recent decades to increase the size of networks accessible by experimental methods, including the new promising macroscale and mesoscale connectivity mapping techniques (Chung et al. 2013; Oh et al. 2014). However, on the microscale of individual neurons, the practical limitations of these techniques mean that reliable reconstruction is currently only possible for neural circuits of up to dozens of cells.
Alternatively, the connectivity of neuronal networks can be inferred from parallel recordings of their spiking activity. Potentially, this enables the recovery of the connections in circuits of hundreds and thousands of cells. Recent technical achievements in conducting largescale parallel recordings of neuronal dynamics, such as multielectrode array technology for in vivo implantation (Hatsopoulos and Donoghue 2009; GhaneMotlagh and Sawan 2013), microelectrode dishes for recording the in vitro activity of acute brain slices and dissociated cell cultures (Nam and Wheeler 2011; Spira and Hai 2013), and optical imaging techniques (Grewe and Helmchen 2009; Lütcke et al. 2013; Ahrens et al. 2013), make this path even more compelling.
The main difficulty in the analysis of parallel recordings, though, lies in the interpretation of the results (Gerstein and Perkel 1969; Aertsen et al. 1989). On one hand, simple reduced models of network interactions are often unable to resolve ambiguous scenarios: a classic example of such ambiguity is a group of neurons that receives common input versus a mutually connected group of cells, which cannot be distinguished using pairwise crosscorrelation analysis (Stevenson et al. 2008). On the other hand, obtaining reliable fits of complex largescale models to the data presents both a methodological and computational challenge in itself (Chen et al. 2011; Song et al. 2013). At the same time, there are often considerable difficulties in directly relating the reconstructed connectivity matrices to measurable experimental quantities or model parameters. The resulting sets of connections are then regarded as “functional” or “effective” connectivity, terms lacking strict and universally accepted definitions, and not necessarily matching real anatomical connectivity, but still hoped to provide useful insights with respect to the interaction of the network elements (Horwitz 2003).
The desire to strike the balance between explanatory power, and analytical as well as numerical tractability, has fueled an ever growing interest in methods that go beyond simple linear regression analysis, but still remain highly efficient. Previous works show that generalized linear models (GLM) (McCullagh and Nelder 1989) of network spiking activity can indeed be efficiently estimated from experimental data (Truccolo et al. 2005; Okatan et al. 2005; Pillow et al. 2008; Stevenson et al. 2009; Gerwinn et al. 2010) (dealing with recordings of up to 20, 33, 27, 75+108 and 7 neurons respectively), and make it possible to recover the actual synaptic connectivity of small neuronal circuits (N=3) (Gerhard et al. 2013). Scaling these approaches directly up to substantially larger networks of thousands of units, however, seemed not to be feasible due to the vast computational resources such a reconstruction would require.
In this work, we present a method to reconstruct the parameters of largescale recurrent neuronal network models of N≥1000 elements, based on parameter estimation of a stochastic point process GLM using only observations of the spiking activity of the neurons. Provided with the knowledge of the probability p(X𝜃) of a specific stochastic model yielding the observations X given the parameters 𝜃, we maximize the likelihood function L(𝜃)=p(X𝜃) in order to identify a set of parameters 𝜃 resulting in an optimal agreement of the selected model with the observations X. This is a widespread technique known as maximum likelihood estimation (MLE) (Paninski 2004). If the underlying model is sufficiently detailed and is indeed appropriate to describe the observations, then not only can the parameters 𝜃 be related to the actual measurable features of the neuronal network that generated the data, but they also define a dynamic model of the neuronal network activity (also called a generative model). Such a model can be used to derive testable predictions, or conduct virtual experiments (simulations), which might otherwise have been impossible or impractical.
Due to the large number of parameters necessary to describe a network of N≥1000 neurons, the optimization of the likelihood L(𝜃) can only be performed efficiently for some of the possible GLMs of neuronal networks. In Section 2, we describe our optimized model, including a particular choice of nonlinearity and interaction kernels, which enables us to obtain closed forms and recurrence formulae which go beyond more general techniques previously reported in the literature. We additionally supply details about the numerical methods employed. In Section 3, we demonstrate the proposed technique on simulations of random balanced neuronal networks, and present reconstructions of the connectivity matrix consisting of 10^{6} possible synapses in sparsely connected recurrent networks of N=1000 spiking neurons. Finally, we apply our method to a structured network. We recover a synfire chain embedded in a balanced network from recordings of spiking activity, in which no activations of the synfire chain were present, and demonstrate that the inferred model of this network supports the transmission of synfire activity when stimulated.
In the present study we focus on reconstructions of networks for which all spiking activity can be recorded. Whereas in experimental settings undersampling is to be expected – and we performed a basic assessment of how it would affect our reconstructions, see Appendix C – a thorough investigation of the consequences of undersampling for the classification performance of our techniques is out of scope. Similarly, when presenting these techniques we are initially concerned with activity which we can assume to be a sample of a multidimensional point process with constant parameters (i.e. neuronal excitability and synaptic interactions). In Section 4 we examine these limitations and propose how they could be relaxed in future studies.
2 Methods
This section provides detailed information on the method of network reconstruction we employ, including original amendments and adaptations. In Section 2.1 we introduce the likelihood of our network model to reproduce a given dataset of neuronal spike trains. This likelihood is the quantity which is subject to optimization. The specific formulation of the likelihood relies on a model of the spiking activity of the neurons, which is introduced in Section 2.2. To evaluate the likelihood and its gradient under that model efficiently, recursive formulae and closed form expressions are derived in Section 2.3. The subsequent sections describe how we handle synaptic transmission delays (Section 2.4) and how, in some cases, we employ regularization of the optimization problem (Section 2.5). Finally, Section 2.6 gives further details regarding the practical aspects of our highly parallelized implementation of the method.
2.1 Point process likelihood of generalized linear models
A statistical model that describes the activity of a network of N neurons can be defined as an expression for the conditional probability \(p(S\vec {x})\) of observing an Ndimensional spike train (spike raster) S for a given input signal \(\vec {x}\), which may include external stimulation and/or previous activity of the network itself. Given all the inputs of a neuron, we assume that its probability of spiking is independent of the other neurons (conditional independence). This allows us to factorize \(p(S\vec {x})={\prod }_{i=1}^{N}p_{i}(S_{i}\vec {x})\), where \(p_{i}(S_{i}\vec {x})\) is the probability that the ith neuron, within the recording time [T _{0},T _{1}], produces a spike train S _{ i } conditioned on the input \(\vec {x}\). Therefore, in what follows we focus on the probability \(p_{i}(S_{i}\vec {x})\) of a single neuron.
The activity of the individual nerve cells can be characterized by a stochastic GLM that postulates that two consecutive operations are performed by the neuron on its input. First, the dimensionality of the observable signal \(\vec {x}\) is reduced by means of a linear transformation K _{ i }. This transformation models synaptic and dendritic filtering, input summation and leaky integration in the soma. The result \(\mathbf {K}_{i}\vec {x}\) is a onedimensional quantity that is analogous to the membrane potential of a point neuron model. Second, this transformed onedimensional signal is fed into a nonlinear probabilistic spiking mechanism, which works by sampling from an inhomogeneous Poisson process with an instantaneous rate (conditional intensity function) given by \(\lambda _{i}(t\vec {x})=f_{i}(\mathbf {K}_{i}\vec {x})\). Here, f _{ i }(⋅) is a function that captures the nonlinear properties of the neuron. Both the linear filter K _{ i } and the nonlinearity f _{ i } are specified by 𝜃 _{ i }, a set of parameters that describes the characteristics of the ith neuron. The schematic of this model is shown in Fig. 1.
Based on these definitions, we may now introduce the natural logarithm \(\mathcal {L}\) of the likelihood L(𝜃S) and expand it as
where the observation (previously called X) is the spike raster S. In the last step of Eq. (1) we have introduced the single neuron loglikelihood \(\mathcal {L}_{i}=\log p_{i}(S_{i}\vec {x})\).
Let us now compute the probability that an inhomogeneous Poisson process with intensity λ _{ i }(t) produces the spike train S _{ i }={t _{ i,k }}, 1≤k≤q _{ i }, where T _{0}≤t _{ i,k }≤T _{1} and q _{ i } is the number of spikes of the ith neuron. This probability is (Brillinger 1988)
with t _{ i,0}=T _{0}. Here, for each spike time t _{ i,k }, we multiply the (survival) probabilities \(e^{{\int }_{\!\!t_{i,k1}}^{t_{i,k}}\lambda _{i}(t)dt}\) of not producing a spike in (t _{ i,k−1},t _{ i,k }) with the intensity λ _{ i }(t _{ i,k }) at t _{ i,k }. Finally, we factor in the probability \(e^{{\int }_{\!\!t_{i,q_{i}}}^{T_{1}}\lambda _{i}(t)dt}\) of not producing a spike in the recording time \((t_{i,q_{i}},T_{1}]\), which remains after the last spike. The function \(L_i (\theta _i  S_i)=p_{i}(S_{i}\vec {x})\) is known as the point process likelihood (Snyder and Miller 1991).
Taking the logarithm yields the loglikelihood function
where the sum runs over all spikes 1≤k≤q _{ i } of the ith neuron. The first term of this expression rewards high intensity at times t _{ i,k } when the spikes of the ith neuron have been emitted, and the second term penalizes high intensity when no spikes have been observed. Different numbers of spikes q _{ i } render the absolute values of \(\mathcal {L}_{i}\) difficult to compare among different neurons, but play no role when maximizing \(\mathcal {L}_{i}\) with respect to 𝜃 _{ i }.
2.2 Conditional intensity model for a recurrent neural network
In order to investigate the recurrent aspects of the dynamics of the system, we define the observable input signal \(\vec {x}\) for each neuron as the history of spikes recorded in the network up to a given point in time, including the spikes of the ith neuron itself (which are used to model the refractory properties of the neuron). It is possible to include external inputs in this formulation, however this is not an option that we have pursued in the current work. Below follows a detailed discussion of the different components of the model as presented in Fig. 1.
For simplicity, we assume that the effect of each incoming spike can be modeled as an instantaneous current injection. The spike train S _{ j } of the jth neuron as a function of time is expressed as \(s_{j}(t)= {\sum }_{k=1}^{q_{j}} \delta (tt_{j,k})\), where t _{ j,k } is the kth spike of the jth neuron. Each spike then elicits an exponential postsynaptic response in the neuron, due to the filtering properties of the membrane, \(h_{i}(t)=H(t)\exp \left \{ {t}/{\tau _{i}}\right \} \), where t is the time since spike arrival, τ _{ i } is the membrane time constant of the neuron, and H(x)={1 if x≥0, else 0} is the Heaviside function, which ensures the causal relationship between the stimulation and the response. Note that while the propagation of spikes is assumed to happen instantaneously in the formulation above, the incorporation of delays will be discussed in detail later in Section 2.4.
We may now define the linear dimensionalityreducing transformation \(U_{i}(t)=\mathbf {K}_{i}\vec {x}(t)\) as
where ∗ denotes the convolution operation,
The baseline potential J _{ i0} will be used later to set a base level of activity of the unit in the absence of inputs. Differentiation of Eq. (3) yields the firstorder ordinary differential equation of the leaky integrator
Hence U _{ i }(t) can be interpreted as the membrane potential of the ith neuron, while J is the synaptic connectivity matrix and each of its elements J _{ i j } denotes the strength (synaptic weight) of the connection from the jth to the ith neuron. Due to its simplicity, Eq. (4) leads to highly efficient algorithms (discussed in Section 2.3 and 2.6) to evaluate the membrane potential and the conditional intensity function of the neurons, beyond previously reported more general parallelization techniques (Chu et al. 2006). The membrane potential and the intensity are, in turn, needed to compute the values of the likelihood function and its gradient.
In Eqs. (3) and (4), positive and negative values of J _{ i j } correspond to excitatory and inhibitory connections respectively, and zero values denote the lack of a connection between two cells. Note that as formulated, this model does not ensure compliance with Dale’s law (according to which each neuron can form synapses of only one type). However, we will show that this is an essentially negligible source of errors in the reconstructions presented below.
Further, we choose a specific type of the nonlinearity \(f(u)=\exp \left \{ u\right \} \), such that
In this expression, the scalar δ u>0 can be considered as the inverse “gain” of the nonlinearity. In the derivations that follow we will assume δ u=1 in order to simplify the expressions without loss of the generality, as different gains can be accommodated by rescaling the synapse weights J _{ i j } and the baseline potential J _{ i0} accordingly. In the absence of input spikes, U _{ i }=J _{ i0}, which leads to the base rate
It is worth mentioning that the base rate can effectively constitute a “sink” for spiking activity that cannot be explained by the recurrent network dynamics, such as external stimulation that has not been included in the present model, or missing inputs from unobserved neurons due to incomplete observations of the network (undersampling).
The model as formulated above is similar to the widely used cascade LNP model (Simoncelli et al. 2004), but in addition to the activity of the other cells in the ensemble, it also incorporates the spiking history of the neuron itself through its selfconnection J _{ i i }. An intuitive biological interpretation of this class of models, also known as the spikeresponse model with escape noise, in relation to the conventional integrateandfire model is given in Brillinger (1988) and Gerstner et al. (2014). Here, in contrast to the approaches taken in previous studies (Song et al. 2013; Citi et al. 2014; Ramirez and Paninski 2014), we drastically simplify both the conditional intensity model for a single neuron and the interaction kernels. This makes the numerics in our method amenable to a highly efficient implementation as discussed in Section 2.6.
Given that f _{ i }(⋅) is both a convex and logconcave function of \(U_{i}=\mathbf {K}_{i}\vec {x}\), and the space of possible {K _{ i }} is convex, it can be shown that the loglikelihood function of such problems is concave and does not have any nonglobal local extrema (Paninski 2004). Thus the loglikelihood function \(\mathcal {L}_{i}\) of the model as formulated above is concave in \(\theta _{i}\subset \{J_{ij}\}_{0\leq j\leq N}\) (note, however, that τ _{ i } is not included in 𝜃 _{ i }; the recovery of the time constants will be addressed separately). A proof of the concavity of \(\mathcal {L}_{i}\) for our specific choice of kernels and link function is given in Appendix A. Since the sum of concave functions is again concave, the full loglikelihood \(\mathcal {L}={\sum }_{i=1}^{N}\mathcal {L}_{i}\) is concave as well. Consequently, there exists a unique set of parameters 𝜃 that characterize the network model that is most likely to exhibit a given recorded activity. These parameters 𝜃 can be efficiently identified via gradient ascent based nonlinear optimization methods applied to \(\mathcal {L}\). Moreover, due to the separability of \(\mathcal {L}\)(1), in order to recover 𝜃={𝜃 _{ i }}, one can maximize the individual loglikelihood functions \(\mathcal {L}_{i}\) for each recorded unit, instead of maximizing the complete loglikelihood function \(\mathcal {L}\).
Since the experimental techniques to obtain simultaneous recordings of thousands of units are becoming increasingly accessible, in this work we are targeting N≥1000. However, even if the number of variables is reduced from the \(\mathcal {\sim O}(N^{2}=10^{6})\) required for the complete loglikelihood function to the \(\mathcal {\sim O}(N=10^{3})\) required for the loglikelihood function of an individual neuron, this is still a highdimensional convex optimization problem. It can only be solved in practice using gradient based methods, for which the analytical closed form expressions for the loglikelihood function and its gradient are both available, and amenable to efficient evaluation. In the following we derive these expressions for the postulated model.
2.3 Closed form expressions
Let us consider the loglikelihood \(\mathcal {L}_{i}\) for an individual neuron; recall that the variable part of Eq. (2) consists of two terms:
Observe that given a closed form for U _{ i }(t), computing \(\mathcal {L}_{i}^{\Sigma }\) is a matter of a simple algebraic substitution, while the efficiency of computing \(\mathcal {L}_{i}^{\int }\) depends on whether it is possible to find this primitive analytically.
2.3.1 Recurrence formula for the membrane potential
By design, our particular choice of K _{ i } (exponential postsynaptic potential plus baseline potential) allows us to obtain the required closed form for U _{ i }(t) because it obeys the leaky integrator dynamics (4). The solution of Eq. (4) from t _{ k } to t in the absence of input spikes s _{ j }(t) is \(U_{i}(t)=(U_{i}(t_{k})J_{i0})\exp \left \{ \frac {tt_{k}}{\tau _{i}}\right \} +J_{i0}\). This expression is valid at any time t between two consecutive observed spikes t _{ k },t _{ k+1}∈S, where S={t _{ k }} is the (ordered) set of all recorded spikes of the network. At the borders of each of those intervals, the value of U _{ i }(t _{ k+1}) is increased by the contribution of the corresponding incoming spike:
where the index j refers to the neuron that emitted a spike at time t _{ k+1}; if spikes from multiple neurons j ^{1,2,3,…} arrive at time t _{ k+1}, the contributions \(J_{ij^{1,2,3,\ldots }}\) have to be added. We will refer to Eq. (8) as the key recurrence formula in the following.
The formula (8) for U _{ i }(t _{ k+1}) makes it possible to find the value of the membrane potential of the neuron at the spike time t _{ k+1} given the previous value at time t _{ k } by computing only one exponential function. It is substantially more efficient in terms of computation than naively summing up the contributions from all spikes that happened at t<t _{ k } for each point in time t _{ k }. In particular, for kernels with infinite memory like the exponential kernels h _{ i }(t) employed here, the recurrence formula (8) is crucial to avoid an explosion of the computational costs when evaluating the loglikelihood on large datasets in continuous time.
2.3.2 Evaluating the likelihood
Taking these considerations into account, the integral over the duration of the recording \(\mathcal {L}_{i}^{\int }\) in Eq. (7) can be broken down into a sum of integrals from t _{ k } to t _{ k+1}:
where \(q={\sum }_{i=1}^{N}q_{i}\) is the total number of recorded spikes, \(t_{1},\dots ,t_{q}\) are the spike times, and t _{0}=T _{0} and t _{ q+1}=T _{1} are the start and end of the recording. The integral contained here has a known closed form, so
where Ei(x) is a special function (exponential integral) defined as \(\text {Ei}(x)={\int }_{\!\!x}^{\infty }\frac {e^{t}}{t}dt\) for real nonzero values of x. For a proof of the equivalence of Eqs. (9) and (10) see Appendix B; the numerical computation of this function is discussed below in Section 2.6.1. The summands of Eq. (10) are independent, and therefore the evaluation of \(\mathcal {L}_{i}^{\int }\) lends itself to trivial parallelization.
2.3.3 Evaluating the gradient
It now remains to find an efficient way to compute the gradient of the loglikelihood function. The performance at this point is likewise important, or even more so for large N, since \(\mathcal {L}_{i}\) has \(\mathcal {O}(N)\) partial derivatives that all need to be evaluated at each step of the optimization. The parameters of \(\mathcal {L}_{i}\) are \(\theta _{i}=(J_{i0},{\dots } J_{iN})\). For convenience, let us first introduce the terms
which, for j≥1, can be interpreted as the putative response of the ith neuron to the input spikes from the jth neuron, that is going to be scaled by J _{ i j }, cf. (3). The derivatives of \(\mathcal {L}_{i}\)(7) with respect to J _{ i j } can then be expressed as
Here, q _{ i } is the number of spikes of the ith neuron, and {t _{ i,k }}=S _{ i } are the points in time when the ith neuron emitted a spike. For j=0, Eq. (12) becomes \(\frac {\partial }{\partial J_{i0}}\mathcal {L}_{i}=q_{i}\mathcal {L}_{i}^{\int }\). This means that at a maximum of \(\mathcal {L}_{i}\), the baseline potential J _{ i0} (and so the base rate c _{ i }(6)) is set such that the number of spikes q _{ i } equals the expected total number of spikes of the GLM, \(\mathcal {L}_{i}^{\int }\). Further, in order to evaluate (12) for the cases when j≥1, we have defined the symbols \(\partial _{ij}^{\Sigma }\) and \(\partial _{ij}^{\int }\) analogous to Eq. (7).
The values ν _{ i j }(t _{ i,k }) for j≥1 can be obtained using a recurrence formula just like for the membrane potential U _{ i }(t _{ k })(8); in fact, \(\nu _{ij}(t)=\frac {\partial }{\partial J_{ij}}U_{i}(t)\), cf. (11). Hence, ν _{ i j }(t) obeys leaky integrator dynamics like U _{ i }(t), which can be obtained by differentiating Eq. (4) by J _{ i j } and reinserting Eq. (11). Accordingly, ν _{ i j }(t) decays exponentially in between spikes \(\nu _{ij}(t_{j,k}<t<t_{j,k+1})=\nu _{ij}(t_{k})\exp \left \{ \frac {tt_{j,k}}{\tau _{i}}\right \} \), and we find the recurrence formula
Individual values within these intervals can be computed in parallel independently from each other. Summing up all ν _{ i j }(t _{ i,k }) then yields \(\partial _{ij}^{\Sigma }\).
It is also important to mention that ν _{ i j }(t)(11) and, consequently, \(\partial _{ij}^{\Sigma }\) in Eq. (12) do not depend on parameters 𝜃 _{ i } and therefore need only be computed once at the beginning of the optimization. However, even though we can use the formula \(U_{i}(t)={\sum }_{j=0}^{N}J_{ij}\nu _{ij}(t)\), for large N it is more expensive to compute U _{ i }(t) by summing up weighted contributions of ν _{ i j }(t) than by using Eq. (8) as explained above.
Making use of the recurrence formulae for U _{ i }(t)(8) and ν _{ i j }(t)(13), the closed form of \(\partial _{ij}^{\int }\) in Eq. (12) can be expressed as follows:
where, as in Eq. (10), \(q={\sum }_{i=1}^{N}q_{i}\) is the total number of recorded spikes, \(t_{1},\dots ,t_{q}\) are the spike times, t _{0}=T _{0} and t _{ q+1}=T _{1} are the start and end of the recording. Unlike \(\partial _{ij}^{\Sigma }\), this expression needs to be reevaluated at every optimization step, but as with Eq. (10), the elements of the sum are independent from each other and can therefore also be efficiently parallelized.
2.4 Handling transmission delays
In the discussion above, the communication of spikes between the neurons was implicitly assumed to happen instantaneously. Of course, in reality spikes incur transmission delays, which strongly affects the dynamics of the network.
Fortunately, the effects of combined synaptic and axonal delays can be easily incorporated into the described model: thanks to the separability property, we can optimize the parameters for each neuron independently, and feed every optimization for different neurons with its own modified dataset, containing the incoming spike times from other neurons arriving as the target neuron actually received them.
Therefore, given an effective delay matrix D, it is only necessary to shift each spike train S _{ j } in the recorded raster S by the corresponding delay at the beginning of the optimization for the ith neuron, such that the membrane potential of this neuron is affected at the point in time when the incoming spikes from the jth neuron have reached their target, and not immediately as they were fired (and recorded):
The transformation above has to be applied with one exception: the elements of the sum in \(\mathcal {L}_{i}^{\Sigma }\) (and, accordingly, \(\partial _{ij}^{\Sigma }\)) have to be evaluated at time points S _{ i } when the ith neuron actually produced a spike, and not at time points \(\widehat {S}{}_{i}=S_{i}+D_{ii}\), when this spike has reached the neuron through the “selfconnection” and provoked a depression of its membrane potential, which models the refractory properties of the neuron.
In other words, in order to correctly evaluate the expressions Eqs. (7) and (12) while taking into account transmission delays, one must compute the values of \(\widehat {U}_{i}(t)\) and \(\widehat {\nu }{}_{ij}(t)\) using the modified raster \(\widehat {S}\), but at time points S _{ i } of the original raster S, and substitute these values in the elements of the sums \(\mathcal {L}_{i}^{\Sigma }\) and \(\partial _{ij}^{\Sigma }\) respectively, instead of summing up the elements taken at times \(\widehat {S}_{i}\). In the following, we omit the “hats” for notational convenience.
2.5 Regularization of the model
Substantial improvements in the quality of the network reconstruction can be achieved if the model presented above is subjected to standard regularization techniques. These techniques enhance the accuracy of the inference procedure by integrating additional prior knowledge about the system into the optimization process (Meinshausen and Bühlmann 2006; Ravikumar et al. 2010). For instance, we can impose box constraints on reasonable values of the synaptic connection matrix J _{ i j } or base rates c _{ i }, and complement this with a choice of more sophisticated methods, such as ℓ _{1} or ℓ _{2} regularization, exploiting assumed sparsity or smoothness of the expected result, respectively (Chen et al. 2011).
In particular, ℓ _{1} regularization (Tibshirani 1996) has a straightforward Bayesian interpretation in our setting: by penalizing the loglikelihood function (2) with the sum of the absolute values of the synaptic weights J _{ i j }, we impose a sparsityinducing Laplace prior on the soughtfor solution, thereby performing a maximum a posteriori (MAP) estimation. Here the strength of the penalty α reflects the firmness of our belief in the sparseness of the network connectivity:
Possible overfitting due to an inadequate choice of the regularization parameter α can be prevented by separating the dataset into two parts to crossvalidate the recovered synaptic weights, and, in the case that the available data is too scarce, more elaborate techniques such as Kfold crossvalidation and other crossvalidation types (Kohavi 1995) can be employed.
2.6 Practical implementation
The mathematical components described above make it possible to reproduce our estimation procedure. However, we found that without employing additional numerical methods, a naive implementation would be way too slow for practical use. In the following we outline the techniques that helped us to boost the optimization speed by many orders of magnitude, bringing the computational requirements to perform estimations of the connectivity for the networks of \(N\sim \mathcal {O}(10^{3})\) neurons into a practical range for plausible amounts of experimental data.
2.6.1 Efficient evaluation
From the computational perspective, a program that performs the parameter estimation would typically consist of a nonlinear optimization routine, which is provided with callback procedures that are repeatedly called in order to evaluate the objective function (2) and its gradient (12) for any given set of parameters. Hence, the cornerstone guiding principle to achieve best performance is to carefully consider the CPU time versus memory consumption tradeoffs, and cache as many values for these callbacks as feasible.
As the values of U _{ i }(t) for S={t _{ k }} (all spikes of the network) are needed in order to evaluate both the loglikelihood function and its gradient, it makes sense to precompute these values at the beginning of the optimization step. Additionally, as previously noted, the values of ν _{ i j }(t) do not depend on the parameters 𝜃, and therefore both ν _{ i j }(t) and \(\partial _{ij}^{\Sigma }\) can be precomputed during the first optimization step, and reused in all subsequent steps. Likewise, it is important to consider the costs of calculating transcendental functions; whereas they might seem negligible at the first sight, the time taken to compute some 10^{10} exponentials every step is considerable. Therefore, precomputing the values of all subexpressions that do not depend on the parameters, and, in particular, \(\xi _{t_{k}}=\exp \left \{ (t_{k+1}/t_{k}){\tau _{i}}\right \} \) is another possibility to save large amounts of CPU time.
In any case, we recommend using iterative profiling in order to select the relevant optimization targets to add each next level of caching, since, as a general rule, the more caches there are, the more complicated and errorprone it is to keep them consistent and up to date with respect to the changes in parameters. Additionally, this avoids the situations when a sizeable amount of work is invested only to gain minor improvements in speed, due to runtime actually being dominated by different code paths than anticipated.
We observed that the optimization algorithms are (unsurprisingly) sensitive to the precision of the evaluation of the objective function and its gradient, and especially to the consistency between the two. Therefore we rejected using numerical approximations to the gradient, such as values computed using the central differences formula, and employed analytically derived expressions instead. We have also found that better precision of the objective function leads to faster convergence. This particularly concerns the accurate approximation of the exponential integral in Eq. (10). In general, finding an efficient method to evaluate Ei(x), which is a crucial part of Eq. (10), poses a significant computational challenge. However, highquality rational approximations exist in the literature (Cody and Thacher 1969), which make it as fast as evaluating loworder polynomials. In our implementation, we rely on the approximations devised by John Maddock using a custom Remez code, which are part of the Boost C++ library.^{Footnote 1} These approximations are not only highly accurate, but also the fastest that are available to us.
2.6.2 Parallelization and distribution
As the sweeping growth of the clock speeds in the last couple of decades seems to have saturated, the focus is increasingly shifting towards increasing parallelism, and nowadays multicore CPUs are a de facto standard, rather than rare marvels. Therefore, suitability for parallelization is becoming a critical feature to discriminate the algorithms that are appropriate for largescale data analysis. In this section we discuss the parallelization strategies applicable to the model described above.
Owing to the separability of the problem, the highest level approach to parallelize the execution of the optimization is to launch several estimations for different neurons in parallel. This results in a perfect scaling for N _{t}≤N, where N _{t} is the number of simultaneously executed hardware threads. This is clearly a very attractive option due to the relative simplicity of implementation, however, its practical applicability is limited by the amount of the available memory per thread, which quickly becomes a bottleneck for larger networks and bigger amounts of data.
A slightly lowerlevel method is to identify independent elements in the formulae that need to be evaluated at every step of the optimization, and divide this work among several threads within one running process. The summands of \(\mathcal {L}_{i}^{\Sigma }\), \(\mathcal {L}_{i}^{\int }\), \(\partial _{ij}^{\Sigma }\) and \(\partial _{ij}^{\int }\) as defined in Eqs. (7), (10), (12) and (14) are all amenable to that kind of processing. This approach is advantageous to utilize all usable threads from within one process, but its scalability is limited by both the amount of the available memory on a single compute node (as above), and the serial part of the computations, which cannot be parallelized. In our model, it is mainly the calculation of the membrane potential U _{ i }(t)(8) and the membrane responses ν _{ i j }(t), because each value in the recurrence formulae depends on the previous one. The membrane responses ν _{ i j }(t) are less of a problem, since they can be precomputed at the beginning of the optimization as explained above, if one is willing to trade memory consumption for performance. Alternatively, ν _{ i j }(t) can be computed in parallel, which can be faster than fetching the results from memory for a very high number of threads and low memory bandwidth.
We have also explored the possibility of distributing the estimation across several compute nodes, which is not only necessary in order to utilize larger numbers of threads than available on one node, but also allows the computation to make use of the additional memory when the problem gets too large to fit into one machine’s RAM. The most straightforward distribution scheme is to designate one process (rank) to perform serial computations required for every optimization step, broadcast the results and parameters to other ranks, have them do their share of the computations, and, finally, collect the results. The biggest advantage of this scheme lies in its ease of implementation: the communication pattern is very clear, and the code can largely remain unchanged except for the need of a few additional functions to distribute and collect the data.
In our implementation, we performed the calculation of the membrane potential U _{ i }(t), the loglikelihood function \(\mathcal {L}_{i}\) and \( \partial \mathcal {L}_{i}/{\partial J_{i0}}\) on Rank 0, and evenly divided the work to compute \( \partial \mathcal {L}_{i}/{\partial J_{ij}}\), j≥1, among all other ranks. This system scales (almost) linearly up to the point when the amount of time needed to perform the computations on Rank 0 exceeds the amount of time it takes to compute the gradient distributed to all other ranks. Since it takes several orders of magnitude more time to calculate \(\mathcal {L}_{i}\) than \( \partial \mathcal {L}_{i}/{\partial J_{ij}}\), we have found that for N=1000 we can easily distribute each single task up to N _{r}=10…20 ranks.
For production estimations, we combined all three approaches outlined above. The highest level of parallelization was left up to the batch system: for each estimation, we generated and submitted the job scripts for every neuron and let the scheduler optimally backfill the queue. The code was run with N _{t}=8…16, depending on the amount of hardware threads available per processor, and N _{r}=10…20, depending on the amount of available memory per processor and the requirements of the particular estimations. For estimations of size N=1000, this hybrid approach allowed us to scale almost linearly up to \(\mathcal {O}(N_{\mathrm {t}}\times N_{\mathrm {r}}\times N=10^{5})\) cores.
In this context, it becomes clear why not only the convexity, but also the separability property of the optimization problem discussed in Section 2.2 is crucial to our model. In a typical estimation, as described in Section 3, 1 hour recording of N=1000 neurons spiking at ∼5 s^{−1} would contain ∼10^{7} spikes, so the intermediary data to be held in RAM during the optimization would need around ∼10^{14}=10×10^{7}×(10^{3})^{2} bytes or 100 TB of storage capacity. This calculation assumes that the main contribution comes from the precomputed matrix of ν _{ i j }(t) vectors of length 10^{7} stored as doubles and disregards all other factors. From our experience, for some N _{r}×N _{t}=10^{5} threads at ∼2 GHz the optimization would take an order of magnitude of 30 minutes of walltime to converge after about a hundred of iterations.
Currently, these requirements can be barely satisfied by booking a complete supercomputer such as JUROPA,^{Footnote 2} and any substantial increase in the number of units, or in the amount of data to be processed will put the problem beyond our reach. However, while the number of parameters of the complete loglikelihood function \(\mathcal {L}\) in our formulation is \(\mathcal {O}(\theta )\sim N^{2}\), thanks to the abovementioned separability property, the number of parameters of \(\mathcal {L}_{i}\) is linear in the number of units, \(\mathcal {O}(\theta _{i})\sim N\). Not only does this present major practical advantages such as easier scheduling of smaller jobs, but it also makes it possible to solve larger problems at all by proportionally trading the execution time for the amount of resources allocated to the optimization process.
2.6.3 Technical realization
Our model was implemented in Python, an increasingly popular language in the field of computational neuroscience. It relies upon the NumPy and SciPy scientific libraries^{Footnote 3} for essential data structures and algorithms. We used Cython^{Footnote 4} in order to bind to the OpenMPparallelized computational kernels, that we extracted and rewrote in C++ for performance reasons, and in order to access the mathematical functions from Boost C++ library. The distribution was implemented using the Python bindings to MPI, mpi4py.^{Footnote 5}
The optimization was performed via the NLopt^{Footnote 6} package by Steven G. Johnson using the lowstorage BroydenFletcherGoldfarbShanno method (Liu and Nocedal 1989) with support for bound constraints (Byrd et al. 1995) implemented by Ladislav Luksan (LBFGSB). We chose to use BFGS instead of the nonlinear conjugate gradient (CG) algorithm, because the former approximates the inverse Hessian matrix of the problem and uses it to steer the search in the parameter space. This results in improved convergence at the cost of higher iteration overhead. Since in our case the computation of the objective function is substantially more expensive, this tradeoff is worthwhile.
As a stopping condition, we used a criterion based on the fractional tolerance of the objective function value. The optimization was terminated if \(\eta ={\Delta }\mathcal {L}/\mathcal {L}\), where \({\Delta }\mathcal {L}\) is the decrease in the function value from one iteration to next, reached the threshold of \(\tilde {\eta }\). The value of \(\tilde {\eta }\) was selected close to the machine epsilon for the double precision floating point type, as requesting even lower tolerance would not yield a more accurate solution; the typical choice was \(\tilde {\eta }\leq 10^{15}\).
It is worth to note that in the case of ℓ _{1} regularized optimizations, it turned out that all gradientbased algorithms we tried were very much affected by the nonsmoothness at zero, introduced by the regularization term in Eq. (16). A thorough review of the existing approaches to address this issue is presented in (Schmidt et al. 2009); we opted for implementing a smooth 𝜖–ℓ _{1} approximation, originally suggested in Lee et al. (2006):
The derivatives of \(\mathcal {L}_{i}\) with respect to J _{ i j }(12) have to be adjusted by addition of \(\alpha J_{ij}/\sqrt {J_{ij}^{2}+\epsilon }\) respectively. We found that this approximation works well in practice for sufficiently small values of 𝜖<10^{−7} and enables us to use the LBFGSB algorithm without modifications. Additionally, we imposed bound constraints on the model parameters as discussed in Section 2.5; typical constraint ranges were J _{ i j }<50 mV for synaptic weights and \(0.001\text { s}^{1}<c_{i}<100\text { s}^{1}\) for base rates. The recordings were truncated to the first and last recorded spikes, T _{0}=t _{1} and T _{1}=t _{ q }, where q is the total number of recorded spikes.
3 Results
We quantified the effectiveness of our suggested method by performing a series of experiments as illustrated in Fig. 2. In these experiments we simulated neuronal networks with known (ground truth) connectivity, and reconstructed the synaptic weight matrix along with the model parameters of these networks on the basis of the recorded spike times. In this way, estimation results could be readily compared to the original connectivity matrix and model parameters. All simulations presented in this section were carried out with the NEural Simulation Tool (NEST) (Gewaltig and Diesmann 2007) and reconstructions were performed using the CPU implementation of the MLE optimizer as described in Section 2. Although the connectivity is sparse in all experiments considered below, we generally use MLE optimization here; only in Section 3.3, which describes the most difficult of the experiments, we also use regularization in order to demonstrate that our computational framework can handle regularized optimization.
In the following subsections, we present the benchmarks of the proposed technique against simulations of a widely used model of a random balanced network (Brunel 2000) and investigate the effect of choosing different neuron and synapse models, first with homogeneous and then with randomly distributed parameters. Finally, we show a successful reconstruction of a specific, nonrandom network, a “synfire chain” embedded in a balanced random network, only from “background” network activity (where the chain was not stimulated). Finally, by stimulating the synfire chain in a simulation of the estimated model, and comparing resulting dynamics to the output of the original network, we highlight the generative aspect of GLM network models. The abbreviations used in the following sections are summarized in Table 1.
3.1 Random balanced network of GLM neurons
As an initial testbed for our method, we selected a random balanced neural network of excitatory and inhibitory neurons in the asynchronous irregular (AI) spiking regime (Brunel 2000). Random networks do not have any particular structural features that can be exploited by the optimizer in order to improve the quality of the reconstruction, and hence in this sense they represent a “worstcase” type of input that is particularly useful for benchmarking purposes. Such networks are commonly studied using the leaky integrateandfire (LIF) neuron model. However, in order to be able to interpret the followup experiments, we first chose to assess the performance of our estimation method under idealized conditions, in which the simulated and estimated neuron and synapse models coincide: the GLM neuron model as described in Section 2 and simple synapses with exponential postsynaptic potentials.
As discussed in Section 2.2, given several conditions that our GLM satisfies, there is a unique maximum likelihood parameter set for the estimated network model (Paninski 2004). In the limit of an infinite amount of spike data used for model estimation and arbitrarily precise calculations, our method is thus bound to recover the true parameters of the simulated model. Hence, testing the method under idealized conditions, but for finite datasets, allows us to distinguish errors that are purely due to the limited length of the observations and restricted machine precision, from those due to a mismatch between the dynamics of the neuron and synapse models used to generate the data, and the dynamics of the models used to reconstruct the network.
The test network consisted of N=1000 GLM neurons with 80 % : 20 % proportion of excitatory to inhibitory neurons (“pp_psc_delta” model in NEST nomenclature, with a base rate c=5 s^{−1}, membrane time constant of τ=20 ms and a resting potential of V _{r}=0 mV). The nonlinearity gain of the neurons was set to δ u=4 mV as in Jolivet et al. (2006), which defines the scaling and units of a single postsynaptic potential via Eq. (5) (δ u=1 as assumed previously in Section 2 for the sake of convenience would make it unitless). Each connection was realized independently with a connection probability of 𝜖=0.2 (ErdősRényi pgraph). The neurons were connected by synapses with exponential postsynaptic potentials with a peak amplitude of J _{e}=1 mV for excitatory and J _{i}=−5 mV for inhibitory synapses, and a transmission delay of d=1.5 ms. A strong inhibitory selfconnection with J _{s}=−25 mV and a transmission delay of d _{s}=Δt was used to model postspike effects. The simulation progressed in time steps of Δt=0.1 ms (resolution) and the simulation time was T=1 hour. The average firing rate of the neurons was ν=4.2 s^{−1}. The recorded spike trains were fed to the estimation method, assuming known values of the time constant τ, the transmission delays d and the delay of the selfconnection d _{s}. The method produced estimates of the synaptic weight matrix J _{ i j } and the base rates {c _{ i }} for all neurons. The original and reconstructed synaptic weight matrix for this experiment are presented in Fig. 2. Throughout this text we refer to {J _{ i j }}_{1≤i,j≤N, i≠j } as the weight matrix; the selfconnections {J _{ i i }}_{1≤i≤N } and the baseline potentials {J _{ i0}}_{1≤i≤N } are treated separately.
In order to evaluate the quality of the reconstruction, we analyzed the resulting distributions of recovered synaptic weights and base rates, as shown in Fig. 3. Whereas the probability density function (PDF) of the original distribution of synaptic weights can be described as a sum of three δfunctions (for excitatory, inhibitory and null connections respectively), the peaks in the reconstructed distribution are broader due to the finite duration of the recording and limited machine precision, to the extent that for realistic values of parameters, there is a degree of overlap between the components of the distributions that represent excitatory and null connections. We noted that the amplitude of the noise that causes the broadening decreases approximately in inverse proportion to the square root of the duration of the recording (data not shown), however, we selected T=1 hour as a reasonable standard amount of input data to mimic conditions where the duration of the recording is limited due to experimental and computational constraints.
This circumstance thus makes it difficult to identify weak excitatory connections unambiguously, and therefore an advanced approach to classification was needed to obtain optimal network reconstruction. To this end, we fitted a Gaussian mixture model (GMM) with a fixed number of components (n=3) to the reconstructed synaptic weights, assuming that synaptic connections, in general, can be either excitatory or inhibitory, or absent. We used an expectationmaximization (EM) algorithm to obtain a maximum likelihood estimate (MLE) of the GMM parameters (mixing weights, means and variances of the individual components), and classified the synaptic weights accordingly. The fitting and classification was performed using a Python implementation of GMM (sklearn.mixture.GMM) provided by the scikitlearn toolkit (Pedregosa et al. 2011). In order to reconstruct the PDFs of the base rates and selfconnections, we used both the histogram function from the NumPy library and the Gaussian kernel density estimation (KDE) code from the SciPy library.
The results are illustrated in Fig. 3, which shows that the means of the distributions were almost perfectly reconstructed and that GMM is indeed an appropriate model for this PDF. The recovered base rates and selfconnection weights are also more or less in agreement with the ground truth values. The detailed classification performance breakdown is presented in Table 2, showing that the classification of synaptic connections is nearly optimal for this dataset (assuming that the cost of making a “false positive” error is equal to the cost of the “false negative” error) and the number of misclassified connections is less than 1%.
3.2 Random balanced network of LIF neurons
Having established the baseline performance in ideal conditions, we designed our next experiment to gauge the influence of mismatch between the neuron and synapse models used to generate the data and those used to reconstruct the network. To this end, we generated data with the commonly used, more complex and realistic LIF neuron model with αshaped postsynaptic currents (PSCs). We then carried out the reconstruction as before assuming our simplified GLM neuron model and synapses with exponential postsynaptic potentials. Another important point is that whereas in the previous experiment we assumed that the membrane time constant τ and transmission delays between the neurons d are known in advance, this is certainly not the case in the laboratory setting, and hence a principled way of estimating these parameters is required in order to analyze real physiological data.
To generate the test data, we wired a network similar to the one described in the previous section, but using a LIF instead of a GLM neuron model. As before, we used N=1000 neurons with 80 % : 20 % ratio of excitatory to inhibitory cells, connection probability of 𝜖=0.2 (each connection was realized independently), transmission delay of d=1.5 ms, simulation resolution of Δt=0.1 ms. Synaptic weights were set to \(\hat {J}_{\mathrm {e/i}}=J_{\mathrm {e/i}}\times w\), with J _{e}=1 mV and J _{i}=−5 mV. The latter (J _{e} and J _{i}) were again interpreted as peak PSP amplitudes, where w=w(τ _{m},τ _{s},C) was the scaling factor (specific to the postsynaptic neuron) selected such that an incoming spike passing through a connection with the synaptic weight of w would evoke a PSP with the maximum amplitude of 1 mV. The parameters of the LIF model (“iaf_psc_alpha” in NEST nomenclature) were chosen as follows: membrane capacitance C=250 pF, membrane time constant τ _{m}=20 ms, synaptic time constant τ _{s}=0.5 ms, refractory time t _{r}=2 ms, firing threshold 𝜃=20 mV, resting potential V _{r}=0 mV and reset to V _{r} after each spike. This time, additional to the synaptic input from other simulated neurons, each neuron received independent Poisson process excitatory inputs at a rate of \(\nu _{\mathrm {e}}=1779\text { s}^{1}\) and inhibitory inputs at \(\nu _{\mathrm {i}}=0.2\times \nu _{\mathrm {e}}=356\text { s}^{1}\). These external inputs represent the influence of neurons that are not part of the simulation, and are necessary to achieve asynchronous and irregular activity as in cortical networks (Brunel 2000). The simulation time was set to T=2 hours and the data was cut into training and validation parts of T _{t}=T _{v}=1 hour as explained below. The average neuron firing rate was ν=4.2 s^{−1}, and so matched the average neuron firing rate of the network of the GLM neurons presented above.
In order to recover the GLM parameters τ and d for this experiment, we applied a crossvalidation procedure. It is important to note that we are not expecting to obtain exactly τ=τ _{m}=20 ms and d=1.5 ms due to mismatch between the LIF with αshaped PSCs and GLM with exponential PSPs models. Instead, we want to recover the optimal parameters τ and d for the GLM model to produce most similar dynamics to the recorded spike trains from the LIF model. We split the available data into a training and a validation dataset, and performed reconstructions for a subset of N _{s}=75 neurons on the training dataset varying one parameter, while keeping the other one fixed. The resulting parameter estimates 𝜃 _{ i } were then used to calculate the loglikelihood function \(\mathcal {L}_{i}\) on the validation dataset. Two different datasets (training and validation) were used in order to ensure that the chosen values of the parameters generalize, and are not specific to the training sample. The validation curves are shown in Fig. 4a, c (the curves for the training dataset look identical); note that they all have an easily identifiable maximum. Subsequently, we averaged the locations of the maxima for all trials and performed another crossvalidation run (Fig. 4d, b) for updated values of the parameters. Repeating this procedure of alternatively fixing one parameter and performing crossvalidation for another one would lead us to a local extremum in the (τ,d) parameter space. However, we opted to stop after only a few iterations because the procedure is computationally expensive, and in order to asses if a suboptimal choice of τ=10 ms and d=1.7 ms would lead to acceptable estimation results.
After determining τ=10 ms and d=1.7 ms through the crossvalidation procedure, we used these values to estimate the connectivity and base rates. The results of the connectivity reconstruction on the training dataset were processed in the same way as in the previous subsection and are presented in Fig. 5, with further details on the classification of synaptic connections in Table 3. We find that the reconstruction quality as defined by classification into the groups of excitatory, inhibitory and null connections closely matches the performance on the ideal dataset analyzed in the previous section, despite the mismatch in models and the suboptimal choice of τ and d. Note that in this experiment, the recovered values of synaptic weights in mV cannot be compared directly to the ones that were used in the simulation which produced the data due to the differences between GLM and LIF models, unlike in the first experiment described in Section 3.1. However, this does not matter for the purposes of classification.
3.3 Random balanced network with distributed parameters
To make the reconstruction task more challenging and to create a more realistic benchmark for our method, we amended the network described in the previous subsection to have different parameters J _{e}, J _{i}, d, τ _{m} and τ _{s} for every neuron and synaptic connection, sampled from uniform distributions around each respective mean value (Table 4), which are the same as in the previous experiment. However, instead of trying to recover the individual values of τ _{ i } for each neuron and d _{ i } for every connection, we decided to investigate whether it would be still possible to make a useful reconstruction assuming identical “mean” values of τ for all neurons and d for all connections. Additional motivation for this choice is in that crossvalidation is a computationally expensive procedure: whereas individual estimation might converge in a matter of minutes, the amount of resources needed to scan a multidimensional parameter grid grows quickly and becomes unmanageable. Therefore, we performed crossvalidation on a subset of neurons as described in the previous subsection, and settled for τ=10 ms and d=1.7 ms again (data not shown).
The estimation results for this dataset are shown in Fig. 6 and Table 5 (left panel and left part of the table respectively). The PDFs of the reconstructed synaptic weights were approximated using Gaussian KDE. Obviously, the individual components of the PDF were distorted, because instead of using optimal values for τ _{ i } and d _{ i }, we used rather arbitrarily chosen fixed values for all neurons and connections. However, more importantly, as the components of the original PDF of synaptic weights were broad distributions rather than δfunctions, the resulting recovered distribution components are strongly nonGaussian. Therefore, in this case the EM procedure for GMM fails to converge to reasonable means and variances, and is no longer a viable choice to perform the classification of connections.
However, instead of engaging in more elaborate statistical modeling to overcome this difficulty, we can take a step back and resort to an unsupervised learning technique called kmeans clustering (which is actually a simplification of GMM). This method rejects the probabilistic assignment of data points to components, and instead makes the assumption that each point belongs to one (and only one) cluster, to the centroid of which it is closest in terms of Euclidean distance. This simplification leads to suboptimal classification when the underlying distributions violate these constraints, but the resulting algorithm is fast and robust.
The Voronoi diagrams for kmeans classification are represented in Fig. 6 as solid lines: the colors show which of the three centroids is closest, in blue, green and red for inhibitory, null and excitatory connections, respectively. By comparing the solid curves and envelope of the colored bars it can be seen that in this case there is a significant overlap between the components contributed by null connections and excitatory connections. Therefore, even the most advanced classification strategies will lead to a substantially higher amount of classification errors than in the previous experiments. The classification data using kmeans is given in Table 5 (left part).
Nevertheless, the situation can still be considerably improved: here, we exploited the sparsity of the synaptic connection matrix by regularizing the GLM estimation with a ℓ _{1} penalty term as explained in Section 2.5. Imposing such a prior on the estimation causes shrinking of the distribution of null connections (Tibshirani 1996) and thus enables better separation between the components. However, the choice of the penalty scaling constant α is arbitrary and so we again availed ourselves of a crossvalidation procedure to determine the optimal value for our dataset.
The results of the reconstruction for a subset of the recorded neurons with different values of α on the training dataset are shown in the left panel of Fig. 7. The right panel depicts the subsequent evaluation of the loglikelihood function on the validation dataset. It is important to note that, for optimal results, this procedure should generally be performed for all neurons, and an individual regularization coefficient should be selected for each of the cells. Instead, in order to save computational resources, we only performed it for a subpopulation of neurons and subsequently selected the same value of α=10 for all cells, which is slightly lower than the average, to prevent excessive connection pruning in neurons with small optimal α.
We performed a full ℓ _{1}regularized GLM estimation using α=10, still fixing the parameters to τ=10 ms and d=1.7 ms, the results of which are presented in Fig. 6, right panel and Table 5, right part. The plot shows that the contribution by null connections indeed shrunk significantly, and thus the amount of classification errors was decreased almost by half. At the same time, for some neurons α=10 turned out to be too strong of a regularization factor, and thus the estimator, in an overzealous attempt to find a sparse solution, set to zero some of the weaker excitatory and inhibitory synapse weights. This can be seen as a secondary peak of the red distribution at the origin. A secondary peak of the blue distribution is also present, but scarcely visible due to scale.
3.4 Synfire chain embedded in a random balanced network
3.4.1 Construction of the network model
In this experiment, we turned to structured networks in order to highlight the generative aspects of the proposed GLM model and demonstrate a potential approach to the interpretation of the recovered connectivity. One specific structure of interest, prominent in the context of cortical networks, is called a “synfire chain” (Abeles 1982). The synfire chain, consisting of consecutively linked and synchronously activated groups of neurons, is a thoroughly studied model of signal propagation in the cortex (Diesmann et al. 1999; Goedeke and Diesmann 2008).
We built a simulation of a random balanced network with an embedded synfire chain, simulated the dynamics of this network and recorded its spiking activity, which we then used as input data for the MLE procedure to infer the parameters of our GLM (no regularization was applied in this experiment, unlike in the last case presented in Section 3.3). However, as would be the case with the experimental recordings, we did not assume that we know the “right” ordering of the neuron identifiers. We therefore subjected the recovered connectivity to a clustering process in order to reveal the trace of the synfire chain in the connection matrix. After identifying the synfire chain in the network, we performed a simulation where we stimulated the discovered first “link” of the chain in the original and reconstructed networks, and observed identical dynamics in both cases.
Similarly to the previous experiments, we first constructed a random balanced network of LIF neurons (N=1000) with 80 % : 20 % proportion of excitatory to inhibitory cells. This time, we used “iaf_psc_delta_canon” model in NEST nomenclature; this model is different from the standard “iaf_psc_delta” and “iaf_psc_alpha” LIF neurons in that the points in time when it emits spikes are not tied to the grid defined by the simulation resolution, but rather are recorded precisely as they occur (Morrison et al. 2007; Hanuschkin et al. 2010). Correspondingly, for the external inputs, we employed the continuous time version of the Poisson generator “poisson_generator_ps”. Since this network model works in continuous time and does not require discretization or binning of the spike data, we wanted to examine the implications of feeding the precise spike times to the MLE of the GLM, as opposed to data binned to Δt=0.1 ms simulation resolution as in the previous experiments. The model parameters were fixed to τ _{m}=20 ms, τ _{r}=2 ms, 𝜃=20 mV, and V _{r}=0 mV. Each neuron was set to receive a fixed number of incoming connections (M _{e}=80 excitatory and M _{i}=20 inhibitory), where the presynaptic neurons were randomly selected (without replacement) from the excitatory and inhibitory populations respectively (implemented as “RandomConvergentConnect” function in NEST). Synaptic weights were set to J _{e}=0.9 mV for excitatory, J _{i}=−4.5 mV for inhibitory connections with a transmission delay of d=1.5 ms. Additional independent Poisson process excitatory inputs were supplied at \(\nu _{\mathrm {e}}=2222\text { s}^{1}\) and inhibitory inputs at \(\nu _{\mathrm {i}}=0.25\times \nu _{\mathrm {e}}=556\text { s}^{1}\).
On top of this “background” network, we selected N _{l}=10 groups (links) of \(N^{\propto }=50\) neurons each (\(N_{\mathrm {e}}^{\propto }=40\) excitatory and \(N_{\mathrm {i}}^{\propto }=10\) inhibitory cells) and connected all \(N_{\mathrm {e}}^{\propto }\) excitatory neurons of every group to each of the \(N^{\propto }\) neurons in the next group with \(J_{\mathrm {e}}^{\propto }=1.4\text { mV}\) excitatory synapses (transmission delay d=1.5 ms). Inhibitory neurons in a link of the chain do not have specific connections to the next link in the chain (Hayon et al. 2004). No neuron in the network was part of more than one group of the synfire chain. This way, we created a “hidden” embedded synfire chain, which receives inputs from the background random network and likewise projects outgoing connections to the background network. When the first group of this structure is stimulated in a coordinated fashion, the chain reliably propagates the excitation from one group to the next until it reaches the last one, and terminates. In the absence of such coordinated stimulation, the synfire chain did not activate, and only “background” activity was observed.
3.4.2 Identification of the synfire chain by connectivity clustering
The complete network was simulated for T=2 hours of biological time and exhibited an average firing rate of ν=1.4 s^{−1}. The synfire chain was not stimulated during the simulation, so the spike train recordings contained no instances of propagating synfire activity. The neuron identifiers were randomly shuffled and the resulting spike raster was fed into the MLE reconstruction procedure.
We reasoned that one of the most generic differentiators between the neurons that belong to various groups (inhibitory neurons and excitatory neurons that are, or are not part of the synfire chain) is the relative strengths of the synapses (both incoming and outgoing connections can be considered). Therefore, we can apply a clustering algorithm to the recovered connectivity matrix to discern between several classes of neurons. However, most algorithms (such as kmeans or GMM, employed in the previous sections) require the desired number of clusters to be set explicitly, either through prior knowledge, or by applying statistical or information theory methods to the data to get an estimation.
To circumvent this problem, we carried out an unsupervised learning technique known as hierarchical clustering. It amounts to iteratively repeating the procedure of looking at the discovered clusters (which, in the first step, each contain a single element), determining the ones that are most similar according to a chosen metric, and merging them into an agglomerate cluster; the process continues until a single cluster remains. The results are visualized by constructing a socalled “dendrogram”, which shows the discovered hierarchy of clusters as a tree structure. Therefore, it is not necessary to specify the number of clusters in advance, but rather the most appropriate set of clusters can be selected by analyzing the dendrogram after performing the clustering. This approach fits very well to an exploratory setting, where one might wish to appreciate the entirety of possible groupings in a compact graphical form and then choose the one that best highlights the particular aspect of interest of the data.
We applied hierarchical clustering to the connectivity matrix using Ward’s minimum variance method (Ward 1963) as a criterion for choosing the pair of clusters to merge at each step. Ward’s minimum variance criterion minimizes the total withincluster variance and enables the grouping of items into sets such that they are maximally similar to each other according to some definition of similarity, which is usually expressed in form of a “dissimilarity matrix”. We used the SciPy hierarchical clustering package (scipy.cluster.hierarchy) to obtain the linkage and visualize the results.
Initially, we grouped the neurons by using the outgoing synaptic weights as the measure of dissimilarity, as shown for the MLEreconstructed connectivity in Fig. 8a. This clustering enabled us to tell excitatory and inhibitory neurons apart (smaller blueish group on the left, and larger reddish group on the right of the matrix). Additionally, in this figure, we can see eight big red squares, which represent the links of the synfire chain. In total, nine squares should be visible in the connectivity matrix for N _{l}=10 links, because the outgoing connections of the last link are not statistically different from those of the background neurons.
The square missing from Fig. 8a is the last link of the chain, which by construction cannot be detected via clustering by the outgoing connections. Therefore, we subjected the neurons that are part of the big yellow cluster (excitatory neurons, which have not been previously identified as taking part in any of the synfire chain links) to additional clustering by incoming connections. This operation reveals the formerly concealed last link of the chain (Fig. 8b). Finally, we applied the same procedure to the inhibitory neurons in the big green cluster. This reveals the inhibitory neurons that are part of the synfire chain. These neurons receive connections from the previous link in the chain but do not send outgoing projections to the next links, and so they are also impossible to detect by clustering only by outgoing connections. This step completes the clustering procedure and we arrive at the final result as shown in Fig. 8c.
In Fig. 9, the clustered matrices (middle column) are contrasted with the matrices in randomized (left column) and original ordering (right column), i.e. the initial indexing of neurons that we used to define the neuron groups of the synfire chain network. An identical clustering procedure was applied to the ground truth connectivity matrix (Fig. 9a–c) and the one obtained from MLE estimation using the recorded spike trains (Fig. 9d–f). Note that, as explained at the end of Section 3.2, the reconstructed values of the synaptic weights in the second row cannot be directly compared to the original synaptic strengths.
The synfire chain is not apparent in the connectivity matrix in randomized ordering, neither for the ground truth matrix (Fig. 9a), nor the MLEestimated connectivity (Fig. 9d). However, clustering neurons by the similarity of incoming and outgoing connection weights reveals the synfire chain substructure (Fig. 9b, e) of both excitatory and inhibitory neurons. This shows that our clustering procedure successfully recovers the group structure of the synfire chain network. Note that in the original ordering (Fig. 9f), the reconstructed matrix also resembles the ground truth matrix to a great extent (Fig. 9c), as expected based on our previous reconstruction experiments above.
3.4.3 Comparison to correlationbased connectivity estimation
In order to compare the results obtained using our GLM method with a wellestablished reference, we also performed lagged crosscorrelation analysis on the same dataset. We computed the crosscorrelation curves ρ _{ i j }(τ) for all pairs of neurons with a bin size of Δ=10 ms and a maximum time lag of \(\tau _{\max }=\pm 200\text { ms}\). The normalized Pearson crosscorrelation coefficient for a stationary ergodic point process for sufficiently large number of sampled bins K is defined as follows (Shao and Chen 1987):
Here, \(\tilde {S_{i}}(t)\) and \(\tilde {S}_{j}(t)\) are binned spike trains of neurons i and j (both Kbins long), whereas N _{ i } and N _{ j } are the total numbers of spikes of the respective neurons. For each ρ(τ) curve, we found the absolute extrema \(\tau _{\text {peak}}^{ij}=\text {argmax}_{\tau }(\rho _{ij}(\tau ))\) and represented the results as a matrix of lagged crosscorrelation coefficients \(\tilde {J}_{ij}=\rho _{ij}(\tau _{\text {peak}}^{ij})\), shown in Fig. 9g–i.
We performed clustering on the \(\tilde {J}_{ij}\) matrix as previously described, however, we had to limit ourselves to the first step only, because the crosscorrelation matrix is symmetric by construction. The matrix shows similar patterns to the ground truth and MLE connectivity matrices, albeit with substantially lower contrast. Additionally, the direction of the synfire chain cannot be detected, due to the symmetry of the measure mentioned above. Moreover, the individual values of the correlation matrix are difficult to directly relate to the experimental quantities and/or model parameters, because the correlation matrix alone does not constitute a generative model, as we discuss below.
3.4.4 Simulation of original and reconstructed synfire chain networks
Finally, we compared the dynamics of the original and reconstructed network in simulation, including occasional stimulation of the first group of the synfire chain. In both networks, we can identify the order of the groups of the chain by following the links backwards starting from the last link identified in Fig. 8b. Note that the identification of the last link is not determined by the clustering algorithm but simply by membership of the neurons as pre and postsynaptic partners in the strong connections represented as red boxes in the clustered connectivity matrix. Neurons in the last link occur just as postsynaptic targets; there is no red box in which they occur as presynaptic sources. Conversely the neurons of the first link only occur as presynaptic sources. Thus the chain can be unrolled from either end by analogous processes.
The simulation of the reconstructed GLM network was carried out in NEST using “pp_psc_delta” neurons and the recovered connectivity matrix. In order to avoid the necessity of finetuning the parameters of the stimulation, we additionally included a reset of the membrane potential \(U_{i}(t)\leftarrow 0\) after spike emission (option “with_reset” in the “pp_psc_delta” model, enabled for all neurons), which prevents runaway excitation of the neurons in the network upon delivering a strong stimulus to the synfire chain.
The results of this experiment are displayed in Fig. 10. We used the order of the neuron identifiers in which the cells were originally wired up to permit a clear visualization of the activity. The raster plots show that the dynamics of the spike patterns of ground truth and estimated network are very similar. Generally, an estimate of a GLM based on recorded spike trains is a generative model of the data, in the sense that, if itself simulated, will produce similar data; Fig. 10 demonstrates this using our embedded synfire chain example.
4 Discussion
In the present work, we introduce a method for analysis of parallel spike trains based upon maximum likelihood estimation of parameters of a recurrent network of stochastic generalized linear model neurons. The method not only makes it possible to perform largescale reconstruction of the directed synaptic connectivity of neuronal circuits, but also to recover neuronal parameters, which can be used to obtain a dynamic (i.e. simulatable) model of the network under investigation. Through radical simplification of the single neuron model and interaction kernels as compared to previous studies (Song et al. 2013; Citi et al. 2014; Ramirez and Paninski 2014), the numerics in our method lend themselves to an efficient implementation on both CPUs and GPUs. Moreover, the estimation procedure is highly amenable to parallelization, which makes it possible to scale up the number of units and putative connections dramatically.
The proposed estimation procedure operates in continuous time on precise timestamps of the events (spikes), and does not require discretization, binning or smoothing of the data, which avoids the associated choice of bin or kernel size and induced artifacts (Ba et al. 2014). Additionally, unlike pairwise methods such as the coupled escape rate model (CERM) by Kobayashi and Kitano (2013), the reconstruction takes into account the complete ensemble spike history and thus is able to disambiguate complex indirect neural interactions. Other recently proposed connectivity reconstruction methods, not based on GLMs, exploit specific properties of leaky integrateandfire neurons (Van Bussel et al. 2011; Memmesheimer et al. 2014) or of linearly interacting point processes (Pernice and Rotter 2013). While this might be less clear for these methods, our procedure, since it is a MLE of a GLM, can be shown to have the optimality properties of becoming an asymptotically unbiased, consistent and efficient estimator of the ground truth connectivity in the limit of large sample sizes (Pawitan 2001) (provided that the suggested model is appropriate to describe neuronal dynamics). Moreover, it is amenable to efficient optimization via gradient ascent, since it is mathematically guaranteed to converge to the global maximum of the likelihood.
We present benchmarks against simulated random balanced networks of N=1000 neurons with known ground truth connectivity, and show that our method achieves good performance for realistic model parameters and plausible amounts of data. Additionally, we performed a successful reconstruction of a structured network, where a synfire chain was embedded in a balanced network of excitatory and inhibitory neurons. The simulation of the reconstructed network with stimulation applied to the first link of the synfire chain, which was identified by carrying out cluster analysis of the recovered synaptic connectivity matrix, highlighted the generative properties of the GLM and showed virtually identical network dynamics to the original network. The application of cluster analysis to the reconstructed connectivity of the synfire chain is an example of how an inferred network model can be subdivided into interacting populations of neurons. Given such a partition of the network in functional subgroups, the activity dynamics can be analyzed using theory of population dynamics of GLM neurons (Deger et al. 2014).
Ideally our approach would also be validated against experimental data. Unfortunately, no datasets are currently available that contain long recordings of many individual spike trains and also the connectivity between the neurons. Indeed, generating such a dataset, although now technically possible, for example, using a high density microelectrode array setup (Ballini et al. 2013), would require extraordinary investment from an experimental laboratory. It is therefore more realistic to hope that the experimental validation of our technique can take place opportunistically on a dataset that is obtained for some other purpose.
In spite of the apparent simplicity of our model, the point process GLM framework that we used is very flexible and can be readily extended with additional features. The exponential kernels that we chose to describe the membrane filtering and nonlinear properties of the neurons can be replaced with more elaborate ones. For instance, previous works have represented neuronal interaction kernels by cosine “bumps” (Pillow et al. 2005), or composition of basis functions, such as Laguerre polynomials or Bsplines (Song et al. 2013). However, in order to enable the reconstruction of networks of thousands of units, the key is to use functions that can both guarantee the concavity of the likelihood, as discussed in Paninski (2004), and at the same time make it possible to find analytical closed forms for the resulting expressions to enable efficient evaluation. These considerations, and the notion that the exponential PSP is a coarse firstorder approximation to the dynamics of synaptic transmission, were the primary motivations for us to adopt the exponential kernels in this work. However, in Section 3.2 we have demonstrated that this simplification does not affect the reconstruction performance for the data generated by a more complex and realistic LIF model with αshaped PSCs. Besides, we argue that our model would be most useful to investigate network effects, as opposed to the effects explained by intricate features of the dynamics of individual synapses, for which purpose, conversely, smallerscale, but more detailed models like the one by Song et al. (2013) might be more appropriate. Apart from that, it is possible to add supplementary terms to the membrane potentials of the neurons U _{ i }(t). One such option is to incorporate known external inputs directly into the model, such as those occurring in experimental paradigms widely used for studying predominantly stimulusdriven circuits like the retina (Pillow et al. 2008). Another option is to add unknown, common external inputs (Kulkarni and Paninski 2007; Vidne et al. 2012) in order to treat nonstationarity in the data.
A further possibility to improve the results of the estimation lies in enforcing Dale’s law: neurons can be either inhibitory or excitatory, and they cannot form connections of both types at the same time (Eccles 1976). Unfortunately, the mathematical reformulation of this law in the context of our model (the sign of all elements in each column of the synaptic weight matrix J _{ i j } should be identical) turns the original problem into a nonconvex and nonseparable one. Instead of trying to solve this much more difficult optimization problem, an approximate, greedy method can be implemented as outlined in Mishchenko et al. (2011). This involves first solving the original problem, then classifying the neurons as excitatory, inhibitory or unassigned based on the discovered synaptic weights, and, finally, imposing corresponding box constraints on the relevant elements of the J _{ i j } rows, which neither compromises the convexity, nor the separability properties. However, in our case, the major source of errors is the overlap between the unconnected and excitatory distributions, which generates nonDale connections as a consequence. The benchmarks that we conducted show that very few of the errors are of the nonDale category (see Tables 2, 3, 5), so any gain from imposing a Dale condition would be minimal and does not justify the additional complexity incurred. Therefore, effort should primarily be focused on tightening distributions.
In our GLM, we have used the exponential link function to map the membrane potential U _{ i }(t) to the instantaneous firing rate λ _{ i }(t). The exponential function is the canonical link function for the Poisson distribution, and it is commonly used in the single neuron modelling context, e.g. in the spike response model (Gerstner et al. 2014). Further reasons for us to choose an exponential function as the link function were as follows: 1) it has been previously shown (Jolivet et al. 2006; Mensi et al. 2012) that an exponential function is a good model for the nonlinear relationship between the conditional intensity of spike emission and the distance from the voltage threshold; 2) an exponential nonlinearity satisfies the sufficient condition established in Paninski (2004) for the likelihood of the model to be concave; 3) this choice makes it possible for us to obtain the closed form for the likelihood function as an exponential integral Ei (10), which is crucial here for reasons of computational efficiency. If the closed form cannot be obtained, then one either needs to discretize the likelihood integral, possibly using clever corrections to improve the accuracy (Citi et al. 2014), or, if the conditional intensity function can be shown to be piecewise smooth like in our case, attempt to get better precision and performance by applying quadrature methods to the smooth segments (Mena and Paninski 2014). Unfortunately, both approaches are still not fast enough for large GLMs such as ours. Other link functions such as logit and probit functions are also commonly used in the context of GLMs and have the property of being bounded (Song et al. 2013). Indeed, within our framework, it is possible to adopt a different link function instead of the canonical one. However, both logit and probit functions in particular are ruled out by the concavity condition (Paninski 2004), being saturating (“sigmoidal”) nonlinearities. In practice, however, we did not experience any substantial issues due to the exponential function being positively unbounded. The box constraints that we imposed on the base rate and synaptic weights served only to repel the optimizer from the borders of the feasible region, where it might occasionally find itself due to an unfortunate combination of numerical artifacts. In none of the results presented in the paper did the recovered parameters turn out to be equal to the values of the box constraints.
Throughout this study, we have assumed that we have simultaneous access to all the spike trains of a neuronal population. For this case, and for a small number of neurons, it was shown that connectivity estimation via GLM can recover anatomical connectivity (Gerhard et al. 2013), as opposed to other methods, which mostly uncover “functional” or “effective” connections (Stevenson et al. 2008) that do not necessarily correspond to real synapses. Here, we scaled the GLM approach up to large networks. However, in many experiments, such as in cortical multielectrode array recordings (Truccolo et al. 2010), a complete recording of all neurons in a network is not feasible, but rather only parts of a neuronal network can be observed. With respect to the inference of connectivity from activity this is known as the problem of undersampling (Kim et al. 2011; Gerhard et al. 2011; Shimazaki et al. 2012; Lütcke et al. 2013): an unobserved neuron might excite several observed ones reliably and frequently. Even if these observed neurons are not synaptically connected to each other, connectivity inference methods that do not account for hidden units would infer connections among them to explain the correlations in their activity. Thus, we generally expect the reconstruction accuracy of our method to decrease in case of undersampling of the network, as the input from unobserved units will be “explained” by nonexisting connections (see also Appendix C.2). Other experimental preparations, such as neuronal cultures on substrateintegrated multielectrode arrays, are amenable to more complete recordings (Ballini et al. 2013), possibly enabling direct interpretation of the recovered connectivity.
We emphasize that our method is practical for networks of up to thousands of neurons, and yet we recognize that the machines featuring a large number of cores (>10^{5}), such as the ones we used during the development phase of this project, are generally only to be found at major research institutions. These supercomputing facilities are becoming increasingly available to neuroscience researchers. For example, researchers based in Germany may take advantage of the twiceyearly calls for applications for computing time on the supercomputers at Jülich Supercomputing Centre^{Footnote 7}, at no cost to the researcher if accepted. European researchers outside Germany can apply analogously for resources through PRACE,^{Footnote 8} and labs based in the US can apply for time at the NSF facilities.^{Footnote 9} Additionally, initiatives such as the Human Brain Project^{Footnote 10} and the Neuroscience Gateway^{Footnote 11} aim to make such resources more accessible to the neuroscience community. Even so, corehour allocations often require a thorough justification and quantitative evidence of the scaling properties of the algorithm, both of which entail significant investment from the researcher in preparing the application.
Therefore, we also investigated the option of offloading the computations to the kinds of GPGPU accelerators that are currently available offtheshelf. We implemented a naive version of a GPU port, in which the computation kernels originally written in C++ and parallelized using OpenMP to use multiple threads were rewritten using CUDA technology by Nvidia Corporation to use a GPU instead. In order to assess the performance of this port, we measured the time it takes to complete the reconstruction of the incoming synapses of one neuron of a network of a thousand of neurons, such as those presented in Section 3. Both applications were tested on an IBM System x iDataPlex dx360 M4 machine featuring two Intel Xeon X5650 processors (6 cores, 12 threads) and one NVIDIA Tesla M2070 (Fermi microarchitecture). The CPU version took 38 minutes to converge in 433 iterations, while the GPU port required 49 minutes and 427 iterations; the obtained loglikelihood values were identical up to an absolute difference of ≃4×10^{−10} and a relative difference of ≃3×10^{−15}. This way, the speedup achieved by offloading the calculations to a single GPU as compared to a single CPU thread amounted to approximately a factor of 18. However, profiling revealed, that around 70 % of the runtime of the GPU port was not actually spent doing useful calculations, but rather transferring ν _{ i j } vectors from the CPU to the GPU memory. Therefore, simply switching to a better GPU, such as the ones based on the Kepler microarchitecture, providing double of the data transfer bandwidth as compared to Fermi, will increase the speedup for a naive GPU port up to a factor of 28. Furthermore, we are currently investigating algorithmic improvements that completely remove the need for data transfers by storing ν _{ i j } vectors directly in the GPU memory using specialized compression. Extrapolating on the performance from the proofofconcept kernels we implemented, a future GPU realization may perform at least as fast as ∼55 generic CPU threads, and require only a fraction of RAM as compared to the CPUonly realization by storing all of the working data in the onboard GPU memory. The complete GPU port of the method, along with its core algorithms and performance benchmarks, will be described in detail in a separate publication. The development of a substantially more efficient implementation will enable us to thoroughly investigate the limits of our approach in a way that is out of scope in the current study due to computational expense. One obvious area for investigation would be the degradation of performance in the case of undersampling as discussed above. Other areas worthy of further examination are the effects on misclassification error rates of correlated external inputs and nonstationarities in the recorded activity.
It is also important to mention that anatomically, cortical neurons receive on the order of 10^{3} 10^{4} incoming synapses (Braitenberg and Schüz 1991). In our demonstrations, we assumed that the network might be fully connected, or, in other words, each of the N=1000 neurons can possibly receive up to 10^{3} incoming synapses from all other neurons, yielding \(\mathcal {O}(N^{2}=10^{6})\) parameters to constrain in total. However, given a substantially larger amount of recording channels, such as N>10^{4}, if such data becomes available, this assumption is no longer reasonable. Instead, the data can be preprocessed to purge unlikely incoming connections, from N down to the most likely 10^{3} 10^{4} putative synapses per neuron, thereby avoiding the quadratic explosion in the number of model parameters. We suggest that such preprocessing can be performed using computationally efficient pairwise linear methods, such as crosscorrelation or crosscoherence, or various information theory metrics (Staniek and Lehnertz 2008). This way, while recovering the connectivity of even larger networks would still require a linear increase in computational resources (or wallclock time), the challenge to further scale the model up to a larger number of putative incoming synapses can be alleviated.
Finally, we would like to stress that even though network models that can be directly simulated as extracted from the data are interesting in themselves for further studies, the proposed method also has potential to provide insights into the networkwide plasticity of synaptic connectivity. Even though in our method we assume that the connectivity is fixed over the time of a recording, estimated synaptic weights can be tracked accross several recordings performed in a timelapse fashion. Such data could be relevant for models of synaptic plasticity over long timescales (structural plasticity) (Escobar et al. 2008; Deger et al. 2012), which currently have to rely on statistics of synapse numbers without temporal information, or timelapse imaging of small numbers of individual synapses.
Notes
References
Abeles, M. (1982). Local cortical circuits: an electrophysiological study, Studies of brain function Vol. 6. Berlin: Springer. doi:10.1007/9783642817083.
Aertsen, A.M.H.J., Gerstein, G.L., Habib, M.K., & Palm, G. (1989). Dynamics of neuronal firing correlation: modulation of “effective connectivity”. Journal of Neurophysiology, 61(5), 900–917.
Ahrens, M.B., Orger, M.B., Robson, D.N., Li, J.M., & Keller, P.J. (2013). Wholebrain functional imaging at cellular resolution using lightsheet microscopy. Nature Methods, 10 (5), 413–420. doi:10.1038/nmeth.2434.
Ba, D., Temereanca, S., & Brown, E.N. (2014). Algorithms for the analysis of ensemble neural spiking activity using simultaneousevent multivariate pointprocess models. Frontiers in Computational Neuroscience, 8, 6. doi:10.3389/fncom.2014.00006.
Ballini, M., Muller, J., Livi, P., Chen, Y., Frey, U., Shadmani, A., Jones, I., Gong, W., Fiscella, M., Radivojevic, M., Bakkum, D., Stettler, A., Heer, F., & Hierlemann, A. (2013). A 1024channel CMOS microelectrodearray system with 26’400 electrodes for recording and stimulation of electroactive cells invitro. In Symposium on VLSI Circuits (VLSIC).
Bock, D.D., Lee, W.C.A., Kerlin, A.M., Andermann, M.L., Hood, G., Wetzel, A.W., Yurgenson, S., Soucy, E.R., Kim, H.S., & Reid, R.C. (2011). Network anatomy and in vivo physiology of visual cortical neurons. Nature, 471(7337), 177–182. doi:10.1038/nature09802.
Boucsein, C., Nawrot, M.P., Schnepel, P., & Aertsen, A. (2011). Beyond the cortical column: abundance and physiology of horizontal connections imply a strong role for inputs from the surround. Frontiers in Neuroscience, 5, 32. doi:10.3389/fnins.2011.00032.
Braitenberg, V., & Schüz, A. (1991). Anatomy of the cortex: Statistics and geometry, Studies of brain function Vol. 18. Berlin: Springer.
Briggman, K.L., Helmstaedter, M., & Denk, W. (2011). Wiring specificity in the directionselectivity circuit of the retina. Nature, 471(7337), 183–188. doi:10.1038/nature09818.
Brillinger, D.R. (1988). Maximum likelihood analysis of spike trains of interacting nerve cells. Biological Cybernetics, 59(3), 189–200.
Brunel, N. (2000). Dynamics of sparsely connected networks of excitatory and inhibitory spiking neurons. Journal of Computational Neuroscience, 8(3), 183–208. doi:10.1023/A:1008925309027.
Byrd, R.H., Lu, P., Nocedal, J., & Zhu, C. (1995). A limited memory algorithm for bound constrained optimization. SIAM Journal on Scientific Computing, 16(5), 1190–1208 . doi:10.1137/0916069.
Chen, Z., Putrino, D.F., Ghosh, S., Barbieri, R., & Brown, E.N. (2011). Statistical inference for assessing functional connectivity of neuronal ensembles with sparse spiking data. IEEE Transactions on Neural Systems and Rehabilitation Engineering, 19(2), 121–135. doi:10.1109/TNSRE.2010.2086079.
Chu, C., Kim, S.K., Lin, Y., Yu, Y., Bradski, G.R., Ng, A.Y., & Olukotun, K. (2006). Mapreduce for machine learning on multicore. In Schölkopf, B., Platt, J., & Hoffman, T. (Eds.) Advances in Neural Information Processing Systems 19, Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems (pp. 281–288). British Columbia, Canada: MIT Press, Vancouver. http://papers.nips.cc/paper/3150mapreduceformachinelearningonmulticore.
Chung, K., Wallace, J., Kim, S.Y., Kalyanasundaram, S., Andalman, A.S., Davidson, T.J., Mirzabekov, J.J., Zalocusky, K.A., Mattis, J., Denisin, A.K., Pak, S., Bernstein, H., Ramakrishnan, C., Grosenick, L., Gradinaru, V., & Deisseroth, K. (2013). Structural and molecular interrogation of intact biological systems. Nature, 497(7449), 332–337. doi:10.1038/nature12107.
Citi, L., Ba, D., Brown, E.N., & Barbieri, R. (2014). Likelihood methods for point processes with refractoriness. Neural Computation, 26(2), 237–263. doi:10.1162/NECO_a_00548.
Cody, W., & Thacher, H.C. (1969). Chebyshev approximations for the exponential integral Ei(x). Mathematics of Computation, 23 (106), 289–303. http://www.ams.org/mcom/196923106/S00255718196902423492/.
Deger, M., Helias, M., Rotter, S., & Diesmann, M. (2012). Spiketiming dependence of structural plasticity explains cooperative synapse formation in the neocortex. PLoS Computational Biology, 8(9), e1002,689. doi:10.1371/journal.pcbi.1002689.
Deger, M., Schwalger, T., Naud, R., & Gerstner, W. (2014). Fluctuations and information filtering in coupled populations of spiking neurons with adaptation. Physical Review E, 90, 062,704. doi:10.1103/PhysRevE.90.062704.
Diesmann, M., Gewaltig, M.O., & Aertsen, A. (1999). Stable propagation of synchronous spiking in cortical neural networks. Nature, 402(6761), 529–533. doi:10.1038/990101.
Eccles, J. (1976). From electrical to chemical transmission in the central nervous system. Notes and Records of the Royal Society of London, 30(2), 219–230.
Escobar, G., Fares, T., & Stepanyants, A. (2008). Structural plasticity of circuits in cortical neuropil. The Journal of Neuroscience, 28(34), 8477–8488. doi:10.1523/JNEUROSCI.204608.2008.
Gerhard, F., Pipa, G., Lima, B., Neuenschwander, S., & Gerstner, W. (2011). Frontiers in Computational Neuroscience, 5, 4. doi:10.3389/fncom.2011.00004.
Gerhard, F., Kispersky, T., Gutierrez, G.J., Marder, E., Kramer, M., & Eden, U. (2013). Successful reconstruction of a physiological circuit with known connectivity from spiking activity alone. PLoS Computational Biology, 9(7), e1003,138. doi:10.1371/journal.pcbi.1003138.
Gerstein, G.L., & Perkel, D.H. (1969). Simultaneously recorded trains of action potentials: analysis and functional interpretation. Science, 164(881), 828–830. doi:10.1126/science.164.3881.828.
Gerstner, W., Kistler, W.M., Naud, R., & Paninski, L. (2014). Neuronal Dynamics: From Single Neurons to Networks and Models of Cognition. New York: Cambridge University Press.
Gerwinn, S., Macke, J.H., & Bethge, M. (2010). Bayesian inference for generalized linear models for spiking neurons. Frontiers in Computational Neuroscience, 4, 12. doi:10.3389/fncom.2010.00012.
Gewaltig, M.O., & Diesmann, M. (2007). NEST (NEural Simulation Tool). Scholarpedia, 2(4), 1430.
GhaneMotlagh, B., & Sawan, M. (2013). Design and implementation challenges of microelectrode arrays: a review. Materials Sciences and Applications, 4(08), 483. doi:10.4236/msa.2013.48059.
Goedeke, S., & Diesmann, M. (2008). The mechanism of synchronization in feedforward neuronal networks. New Journal of Physics, 10(1), 015,007. doi:10.1088/13672630/10/1/015007.
Grewe, B.F., & Helmchen, F. (2009). Optical probing of neuronal ensemble activity. Curr. Opin. Neurobiol., 19(5), 520–529. doi:10.1016/j.conb.2009.09.003.
Hanuschkin, A., Kunkel, S., Helias, M., Morrison, A., & Diesmann, M. (2010). A general and efficient method for incorporating precise spike times in globally timedriven simulations. Front. Neuroinform., 4, 113. doi:10.3389/fninf.2010.00113.
Hatsopoulos, N.G., & Donoghue, J.P. (2009). The science of neural interface systems. Annual Review of Neuroscience, 32, 249–266. doi:10.1146/annurev.neuro.051508.135241.
Hayon, G., Abeles, M., & Lehmann, D. (2004). Modeling compositionality by dynamic binding of synfire chains. Journal of Computational Neuroscience, 17(2), 179–201. doi:10.1023/B:JCNS.0000037682.18051.5f.
Horwitz, B. (2003). The elusive concept of brain connectivity. Neuroimage, 19(2 Pt 1), 466–470. doi:10.1016/S10538119(03)001125.
Jolivet, R., Rauch, A., Lüscher, H.R., & Gerstner, W. (2006). Predicting spike timing of neocortical pyramidal neurons by simple threshold models. Journal of Computational Neuroscience, 21(1), 35–49. doi:10.1007/s1082700670745.
Kim, S., Putrino, D., Ghosh, S., & Brown, E.N. (2011). A Granger causality measure for point process models of ensemble neural spiking activity. PLoS Computational Biology, 7(3), e1001,110. doi:10.1371/journal.pcbi.1001110.
Kobayashi, R., & Kitano, K. (2013). Impact of network topology on inference of synaptic connectivity from multineuronal spike data simulated by a largescale cortical network model. Journal of Computational Neuroscience, 35(1), 109–124. doi:10.1007/s108270130443y.
Kohavi, R. (1995). A study of crossvalidation and bootstrap for accuracy estimation and model selection. In International Joint Conference on Artificial Intelligence, (Vol. 14 pp. 1137–1145).
Kulkarni, J.E., & Paninski, L. (2007). Commoninput models for multiple neural spiketrain data. Network, 18(4), 375–407. doi:10.1080/09548980701625173.
Lee, S.I., Lee, H., Abbeel, P., & Ng, A.Y. (2006). Efficient ℓ _{1} regularized logistic regression. In Proceedings of the National Conference on Artificial Intelligence, (Vol. 21 p. 401). Menlo Park, CA; Cambridge, MA; London: AAAI Press; MIT Press; 1999.
Liu, D.C., & Nocedal, J. (1989). On the limited memory BFGS method for large scale optimization. Mathematical Programming, 45(13), 503–528. doi:10.1007/bf01589116.
Lütcke, H., Gerhard, F., Zenke, F., Gerstner, W., & Helmchen, F. (2013). Inference of neuronal network spike dynamics and topology from calcium imaging data. Front. Neural. Circuits, 7, 201. doi:10.3389/fncir.2013.00201.
McCullagh, P., & Nelder, J.A. (1989). Generalized linear models, 2nd edn. No. 37 in Monographs on statistics and applied probability. London: Chapman and Hall.
Meinshausen, N., & Bühlmann, P. (2006). Highdimensional graphs and variable selection with the Lasso. The Annals of Statistics, 34(3), 1436–1462. doi:10.1214/009053606000000281.
Memmesheimer, R.M., Rubin, R., Olveczky, B.P., & Sompolinsky, H. (2014). Learning precisely timed spikes. Neuron, 82(4), 925–938. doi:10.1016/j.neuron.2014.03.026.
Mena, G., & Paninski, L. (2014). On quadrature methods for refractory point process likelihoods. Neural Computation, 26(12), 2790–2797. doi:10.1162/NECO_a_00676.
Mensi, S., Naud, R., Pozzorini, C., Avermann, M., Petersen, C.C.H., & Gerstner, W. (2012). Parameter extraction and classification of three cortical neuron types reveals two distinct adaptation mechanisms. Journal of Neurophysiology, 107(6), 1756–1775. doi:10.1152/jn.00408.2011.
Mishchenko, Y., Vogelstein, J., & Paninski, L. (2011). A Bayesian approach for inferring neuronal connectivity from calcium fluorescent imaging data. Annals of Applied Statistics, 5(2B), 1229–1261. doi:10.1214/09AOAS303.
Morrison, A., Straube, S., Plesser, H.E., & Diesmann, M. (2007). Exact subthreshold integration with continuous spike times in discretetime neural network simulations. Neural Computation, 19(1), 47–79. doi:10.1162/neco.2007.19.1.47.
Nam, Y., & Wheeler, B.C. (2011). In vitro microelectrode array technology and neural recordings. Critical Reviews in Biomedical Engineering, 39(1), 45–61. doi:10.1615/CritRevBiomedEng.v39.i1.40.
Oh, S.W., Harris, J.A., Ng, L., Winslow, B., Cain, N., Mihalas, S., Wang, Q., Lau, C., Kuan, L., Henry, A.M., Mortrud, M.T., Ouellette, B., Nguyen, T.N., Sorensen, S.A., Slaughterbeck, C.R., Wakeman, W., Li, Y., Feng, D., Ho, A., Nicholas, E., Hirokawa, K.E., Bohn, P., Joines, K.M., Peng, H., Hawrylycz, M.J., Phillips, J.W., Hohmann, J.G., Wohnoutka, P., Gerfen, C.R., Koch, C., Bernard, A., Dang, C., Jones, A.R., & Zeng, H. (2014). A mesoscale connectome of the mouse brain. Nature, 508(7495), 207–214. doi:10.1038/nature13186.
Okatan, M., Wilson, M.A., & Brown, E.N. (2005). Analyzing functional connectivity using a network likelihood model of ensemble neural spiking activity. Neural Computation, 17(9), 1927–1961. doi:10.1162/0899766054322973.
Paninski, L. (2004). Maximum likelihood estimation of cascade pointprocess neural encoding models. Network, 15(4), 243–262.
Pawitan, Y. (2001). In all likelihood: statistical modelling and inference using likelihood. London: Oxford University Press.
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cournapeau, D., Brucher, M., Perrot, M., & Duchesnay, E. (2011). Scikitlearn: Machine learning in Python. Journal of Machine Learning Research, 12, 2825–2830.
Perin, R., Berger, T.K., & Markram, H. (2011). A synaptic organizing principle for cortical neuronal groups. Proceedings of the National Academy of Sciences of the United States of America, 108(13), 5419–5424. doi:10.1073/pnas.1016051108.
Pernice, V., & Rotter, S. (2013). Reconstruction of sparse connectivity in neural networks from spike train covariances. Journal of Statistical Mechanics: Theory and Experiment, 2013 (03), P03,008. doi:10.1088/17425468/2013/03/p03008.
Pillow, J.W., Paninski, L., Uzzell, V.J., Simoncelli, E.P., & Chichilnisky, E.J. (2005). Prediction and decoding of retinal ganglion cell responses with a probabilistic spiking model. Journal of Neuroscience, 2547, 11,003–11,013. doi:10.1523/JNEUROSCI.330505.2005.
Pillow, J.W., Shlens, J., Paninski, L., Sher, A., Litke, A.M., Chichilnisky, E.J., & Simoncelli, E.P. (2008). Spatiotemporal correlations and visual signalling in a complete neuronal population. Nature, 454(7207), 995–999. doi:10.1038/nature07140.
Ramirez, A.D., & Paninski, L. (2014). Fast inference in generalized linear models via expected loglikelihoods. Journal of Computational Neuroscience, 36(2), 215–234. doi:10.1007/s1082701304664.
Ravikumar, P., Wainwright, M.J., & Lafferty, J.D. (2010). Highdimensional Ising model selection using ℓ _{1}regularized logistic regression. Annal of Statistics, 38(3), 1287–1319. doi:10.1214/09AOS691.
Schmidt, M., Fung, G., & Rosales, R. (2009). Optimization methods for ℓ _{1}regularization. Tech. Rep. TR200919, University of British Columbia. http://www.cs.ubc.ca/cgibin/tr/2009/TR200919.
Shao, X.S., & Chen, P.X. (1987). Normalized auto and crosscovariance functions for neuronal spike train analysis. The International Journal of Neuroscience, 34(12), 85–95.
Shimazaki, H., Amari, S.I., Brown, E.N., & Grün, S. (2012). Statespace analysis of timevarying higherorder spike correlation for multiple neural spike train data. PLoS Computational Biology, 83, e1002,385. doi:10.1371/journal.pcbi.1002385.
Simoncelli, E., Paninski, L., Pillow, J., & Schwartz, O. (2004). The New Cognitive Neurosciences. Cambridge, MA, London, England: MIT Press. chap Characterization of neural responses with stochastic stimuli. Bradford Books.
Snyder, D.L., & Miller, M.I. (1991). Random Point Processes in Time and Space (Springer Texts in Electrical Engineering). New York: Springer. doi:10.1007/9781461231660.
Song, D., Wang, H., Tu, C.Y., Marmarelis, V.Z., Hampson, R.E., Deadwyler, S.A., & Berger, T.W. (2013). Identification of sparse neural functional connectivity using penalized likelihood estimation and basis functions. Journal of Computational Neuroscience, 35(3), 335–357. doi:10.1007/s1082701304557.
Spira, M.E., & Hai, A. (2013). Multielectrode array technologies for neuroscience and cardiology. Nature Nanotechnology, 8(2), 83–94. doi:10.1038/nnano.2012.265.
Staniek, M., & Lehnertz, K. (2008). Symbolic transfer entropy. Physical Review Letters, 100(15), 158,101. doi:10.1103/PhysRevLett.100.158101.
Stevenson, I.H., Rebesco, J.M., Miller, L.E., & Körding, K.P. (2008). Inferring functional connections between neurons. Current Opinion in Neurobiology, 18(6), 582–588. doi:10.1016/j.conb.2008.11.005.
Stevenson, I.H., Rebesco, J.M., Hatsopoulos, N.G., Haga, Z., Miller, L.E., & Körding, K.P. (2009). Bayesian inference of functional connectivity and network structure from spikes. IEEE Transactions on Neural Systems and Rehabilitation, 17(3), 203–213. doi:10.1109/TNSRE.2008.2010471.
Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. J Roy Stat Soc B Met, 267–288.
Truccolo, W., Eden, U.T., Fellows, M.R., Donoghue, J.P., & Brown, E.N. (2005). A point process framework for relating neural spiking activity to spiking history, neural ensemble, and extrinsic covariate effects. Journal of Neurophysiology, 93(2), 1074–1089. doi:10.1152/jn.00697.2004.
Truccolo, W., Hochberg, L.R., & Donoghue, J.P. (2010). Collective dynamics in human and monkey sensorimotor cortex: predicting single neuron spikes. Nature Neuroscience, 13(1), 105–111. doi:10.1038/nn.2455.
Van Bussel, F., Kriener, B., & Timme, M. (2011). Inferring synaptic connectivity from spatiotemporal spike patterns. Frontiers in Computational Neuroscience, 5, 3. doi:10.3389/fncom.2011.00003.
Vidne, M., Ahmadian, Y., Shlens, J., Pillow, J.W., Kulkarni, J., Litke, A.M., Chichilnisky, E.J., Simoncelli, E., & Paninski, L. (2012). Modeling the impact of common noise inputs on the network activity of retinal ganglion cells. Journal of Computational Neuroscience, 33(1), 97–121 . doi:10.1007/s1082701103762.
Ward, J.H. (1963). Hierarchical grouping to optimize an objective function. JJournal of the American Statistical Association, 58(301), 236–244.
Acknowledgments
We thank Stefano Cardanobile, Stefan Rotter, Wolfram Schenck and Christian Pozzorini for helpful discussions, and Ulrich Egert for his support of the initial phase of the project. We are grateful to Andrew V. Adinetz for his advice on analyzing the performance of our optimizer and on the GPU kernel implementation, and to Tom Tetzlaff for providing an example synfire chain simulation script. We acknowledge partial support by the German Federal Ministry of Education and Research (BMBF grants 01GQ0420 to BCCN Freiburg and 01GQ0830 to BFNT Freiburg), the Swiss National Science Foundation (grant agreement no. 200020_147200), and the Helmholtz Alliance through the Initiative and Networking Fund of the Helmholtz Association and the Helmholtz Portfolio theme “Supercomputing and Modeling for the Human Brain”.
Conflict of interests
The authors declare that they have no conflict of interest.
Open Access
This article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.
Author information
Authors and Affiliations
Corresponding author
Additional information
Action Editor: Abraham Zvi Snyder
Appendices
Appendix
The spike data used in this paper and the code that implements our connectivity reconstruction method are publicly available for download at doi:10.5281/zenodo.17662 and doi:10.5281/zenodo.17663 respectively.
A: Concavity of the point process loglikelihood
The derivative of the loglikelihood \(\mathcal {L}_{i}\)(2) with respect to J _{ i j } is
cf. (12), with ν _{ i j }(t)=(h _{ i }∗s _{ j })(t) and ν _{ i0}=1(11). The second derivative is then simply
Using these expressions, in the following we give a proof, specific to our model, that \(\mathcal {L}_{i}\) is concave. A condition for the concavity of the loglikelihood of more general point process GLMs is given in Paninski (2004).
A twice differentiable function of several variables is concave if and only if its Hessian matrix \(\mathcal {H}\) is negative semidefinite. In terms of the parameter vector \(\theta _{i}=(J_{i0},{\dots } J_{iN})\), the Hessian matrix of \(\mathcal {L}_{i}\) has the elements (18). This matrix is negative semidefinite if \(x^{T}\mathcal {H}x\leq 0\) for all real vectors x. We evaluate this expression as
where V _{ i }(t) is defined analogously to U _{ i }(t)(3) as \({\sum }_{j=0}^{N}x_{j}\nu _{ij}(t)\), and λ _{ i }(t)≥0. Hence \(\mathcal {H}\) is negative semidefinite, and thus \(\mathcal {L}_{i}\) is concave in the parameters 𝜃 _{ i }.
B: Closed form for the loglikelihood integral
To evaluate Eq. (9), we need to compute the term
Let us introduce the shorthand \(g(t)=(U_{i}(t_{k})J_{i0})e^{\frac {tt_{k}}{\tau _{i}}}\). We need to show that the exponential integral \(\text {Ei}(x)={\int }_{\!\!x}^{\infty }\frac {e^{t}}{t}dt\) is a primitive of \(\exp \{g(t)\}\) for t _{ k }≤t<t _{ k+1}. We differentiate
Thus, we can evaluate the integral as \({\int }_{\!\!t_{k}}^{t_{k+1}}\exp \left \{ g(t)\right \} dt=\tau _{i}\left .\text {Ei}(g(t))\right _{t_{k}}^{t_{k+1}}\), and so follows (10).
C: Spot checks for several degrees of undersampling and sparsity
1.1 C.1: Chance level of the misclassification error rate
Connections in our networks are formed with a connection probability p. A fraction f _{e} of neurons is excitatory, the remainder f _{i}=1−f _{e} is inhibitory. To assess the quality of our connectivity reconstruction, here we compute the misclassification error rate (MER) of a random connection classifier that maintains p, f _{e} and f _{i}. We call this the chance level MER_{0}.
Misclassification errors can occur for three types of synaptic connections: excitatory, inhibitory and null connections. Let us take the example of the excitatory type. We expect p(f _{e} N)(N−1) excitatory connections, each of which is misclassified (false negative) with probability (1−p f _{e}), because with probability p f _{e} it would be classified correctly as excitatory. Analogously the expected number of misclassifications of each type is given by

excitatory: n _{e}=p(f _{e} N)(N−1)(1−p f _{e});

inhibitory: n _{i}=p(f _{i} N)(N−1)(1−p f _{i});

null: n _{n}=(1−p)N(N−1)p .
The total rate of errors is then the expected number of errors, summed over types, divided by the total number of possible connections. This yields the following expression:
which is independent of N, but depends on the connection probability p and the ratio of excitatory to inhibitory neurons.
1.1.1 C.2: Effects of undersampling
To assess the degree to which undersampling deteriorates the quality of the network reconstructions, we performed several experiments with different datasets, each being a subsample of the original one presented in Section 3.1. In each experiment we randomly selected a fraction of neurons (maintaining the ratio of excitatory and inhibitory neurons) that are fed into the optimizer. The results are shown in Fig. 11. In contrast to Fig. 3, here the connections were classified using kmeans as described in Section 3.3, which is more robust in the undersampled cases. Therefore, for the case of N=1000 neurons the MER is slightly higher than when classified using GMM, as reported in Table 2. As expected, the MER of the partial network increased as we decreased the number of neurons that were visible to the GLM (undersampling). This was largely due to the broadening of the distribution of the synaptic weights for null connections (data not shown, cf. Figure 3). Yet, in all cases, synapse classification based on the reconstruction method was substantially better than random classification of the synapses, see Appendix C.1 for the derivation of the chance level MER.
1.1.2 C.3: Effects of varying connection sparsity
In this experiment we performed the reconstruction on a dataset simulated as described in Section 3.1, but with connection probability of p=0.1 instead of p=0.2. The results are presented in Table 6. Note that whereas the quality of the reconstruction is substantially better than for p=0.2 (shown in Table 2), the chance level of the misclassification error rate for this network with p=0.1 is MER_{0}=18.32 %, rather than MER_{0}=33.28 % for the network with p=0.2. Still, also in relative terms to MER_{0}, the reconstruction is more accurate in this case of increased connection sparsity.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0), which permits use, duplication, adaptation, distribution, and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Zaytsev, Y.V., Morrison, A. & Deger, M. Reconstruction of recurrent synaptic connectivity of thousands of neurons from simulated spiking activity. J Comput Neurosci 39, 77–103 (2015). https://doi.org/10.1007/s1082701505655
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s1082701505655