# Reconstruction of recurrent synaptic connectivity of thousands of neurons from simulated spiking activity

- 2.1k Downloads
- 15 Citations

## Abstract

Dynamics and function of neuronal networks are determined by their synaptic connectivity. Current experimental methods to analyze synaptic network structure on the cellular level, however, cover only small fractions of functional neuronal circuits, typically without a simultaneous record of neuronal spiking activity. Here we present a method for the reconstruction of large recurrent neuronal networks from thousands of parallel spike train recordings. We employ maximum likelihood estimation of a generalized linear model of the spiking activity in continuous time. For this model the point process likelihood is concave, such that a global optimum of the parameters can be obtained by gradient ascent. Previous methods, including those of the same class, did not allow recurrent networks of that order of magnitude to be reconstructed due to prohibitive computational cost and numerical instabilities. We describe a minimal model that is optimized for large networks and an efficient scheme for its parallelized numerical optimization on generic computing clusters. For a simulated balanced random network of 1000 neurons, synaptic connectivity is recovered with a misclassification error rate of less than 1 % under ideal conditions. We show that the error rate remains low in a series of example cases under progressively less ideal conditions. Finally, we successfully reconstruct the connectivity of a hidden synfire chain that is embedded in a random network, which requires clustering of the network connectivity to reveal the synfire groups. Our results demonstrate how synaptic connectivity could potentially be inferred from large-scale parallel spike train recordings.

## Keywords

Spike trains Network topology Connectome identification Inverse problem Synaptic connectivity Connectivity inference Generalized linear model Maximum likelihood estimation Penalized likelihood Sparsity Point process## 1 Introduction

The synaptic organization of neuronal networks is key to understanding the dynamics of brain circuits, and, eventually, to link them to higher level cognitive functions. A large body of work aims to address this challenge by developing experimental techniques which enable the reconstruction of the connections between neurons on the basis of anatomical or physiological evidence. Anatomically, synaptic connections may be identified using optical imaging or electron microscopy (Briggman et al. 2011; Bock et al. 2011), while physiological approaches rely on simultaneous recordings of individual neurons and the mutual influence of the spikes of one neuron on the membrane potential of the other (Perin et al. 2011; Boucsein et al. 2011). Substantial progress has been made in recent decades to increase the size of networks accessible by experimental methods, including the new promising macroscale and mesoscale connectivity mapping techniques (Chung et al. 2013; Oh et al. 2014). However, on the microscale of individual neurons, the practical limitations of these techniques mean that reliable reconstruction is currently only possible for neural circuits of up to dozens of cells.

Alternatively, the connectivity of neuronal networks can be inferred from parallel recordings of their spiking activity. Potentially, this enables the recovery of the connections in circuits of hundreds and thousands of cells. Recent technical achievements in conducting large-scale parallel recordings of neuronal dynamics, such as multi-electrode array technology for *in vivo* implantation (Hatsopoulos and Donoghue 2009; Ghane-Motlagh and Sawan 2013), micro-electrode dishes for recording the *in vitro* activity of acute brain slices and dissociated cell cultures (Nam and Wheeler 2011; Spira and Hai 2013), and optical imaging techniques (Grewe and Helmchen 2009; Lütcke et al. 2013; Ahrens et al. 2013), make this path even more compelling.

The main difficulty in the analysis of parallel recordings, though, lies in the interpretation of the results (Gerstein and Perkel 1969; Aertsen et al. 1989). On one hand, simple reduced models of network interactions are often unable to resolve ambiguous scenarios: a classic example of such ambiguity is a group of neurons that receives common input versus a mutually connected group of cells, which cannot be distinguished using pairwise cross-correlation analysis (Stevenson et al. 2008). On the other hand, obtaining reliable fits of complex large-scale models to the data presents both a methodological and computational challenge in itself (Chen et al. 2011; Song et al. 2013). At the same time, there are often considerable difficulties in directly relating the reconstructed connectivity matrices to measurable experimental quantities or model parameters. The resulting sets of connections are then regarded as “functional” or “effective” connectivity, terms lacking strict and universally accepted definitions, and not necessarily matching real anatomical connectivity, but still hoped to provide useful insights with respect to the interaction of the network elements (Horwitz 2003).

The desire to strike the balance between explanatory power, and analytical as well as numerical tractability, has fueled an ever growing interest in methods that go beyond simple linear regression analysis, but still remain highly efficient. Previous works show that *generalized linear models* (GLM) (McCullagh and Nelder 1989) of network spiking activity can indeed be efficiently estimated from experimental data (Truccolo et al. 2005; Okatan et al. 2005; Pillow et al. 2008; Stevenson et al. 2009; Gerwinn et al. 2010) (dealing with recordings of up to 20, 33, 27, 75+108 and 7 neurons respectively), and make it possible to recover the actual synaptic connectivity of small neuronal circuits (*N*=3) (Gerhard et al. 2013). Scaling these approaches directly up to substantially larger networks of thousands of units, however, seemed not to be feasible due to the vast computational resources such a reconstruction would require.

In this work, we present a method to reconstruct the parameters of large-scale recurrent neuronal network models of *N*≥1000 elements, based on parameter estimation of a stochastic point process GLM using only observations of the spiking activity of the neurons. Provided with the knowledge of the probability *p*(*X*|*𝜃*) of a specific stochastic model yielding the observations *X* given the parameters *𝜃*, we maximize the likelihood function *L*(*𝜃*)=*p*(*X*|*𝜃*) in order to identify a set of parameters *𝜃* resulting in an optimal agreement of the selected model with the observations *X*. This is a widespread technique known as *maximum likelihood estimation* (MLE) (Paninski 2004). If the underlying model is sufficiently detailed and is indeed appropriate to describe the observations, then not only can the parameters *𝜃* be related to the actual measurable features of the neuronal network that generated the data, but they also define a dynamic model of the neuronal network activity (also called a generative model). Such a model can be used to derive testable predictions, or conduct virtual experiments (simulations), which might otherwise have been impossible or impractical.

Due to the large number of parameters necessary to describe a network of *N*≥1000 neurons, the optimization of the likelihood *L*(*𝜃*) can only be performed efficiently for some of the possible GLMs of neuronal networks. In Section 2, we describe our optimized model, including a particular choice of nonlinearity and interaction kernels, which enables us to obtain closed forms and recurrence formulae which go beyond more general techniques previously reported in the literature. We additionally supply details about the numerical methods employed. In Section 3, we demonstrate the proposed technique on simulations of random balanced neuronal networks, and present reconstructions of the connectivity matrix consisting of 10^{6} possible synapses in sparsely connected recurrent networks of *N*=1000 spiking neurons. Finally, we apply our method to a structured network. We recover a synfire chain embedded in a balanced network from recordings of spiking activity, in which no activations of the synfire chain were present, and demonstrate that the inferred model of this network supports the transmission of synfire activity when stimulated.

In the present study we focus on reconstructions of networks for which all spiking activity can be recorded. Whereas in experimental settings undersampling is to be expected – and we performed a basic assessment of how it would affect our reconstructions, see Appendix C – a thorough investigation of the consequences of undersampling for the classification performance of our techniques is out of scope. Similarly, when presenting these techniques we are initially concerned with activity which we can assume to be a sample of a multi-dimensional point process with constant parameters (i.e. neuronal excitability and synaptic interactions). In Section 4 we examine these limitations and propose how they could be relaxed in future studies.

## 2 Methods

This section provides detailed information on the method of network reconstruction we employ, including original amendments and adaptations. In Section 2.1 we introduce the likelihood of our network model to reproduce a given dataset of neuronal spike trains. This likelihood is the quantity which is subject to optimization. The specific formulation of the likelihood relies on a model of the spiking activity of the neurons, which is introduced in Section 2.2. To evaluate the likelihood and its gradient under that model efficiently, recursive formulae and closed form expressions are derived in Section 2.3. The subsequent sections describe how we handle synaptic transmission delays (Section 2.4) and how, in some cases, we employ regularization of the optimization problem (Section 2.5). Finally, Section 2.6 gives further details regarding the practical aspects of our highly parallelized implementation of the method.

### 2.1 Point process likelihood of generalized linear models

A statistical model that describes the activity of a network of *N* neurons can be defined as an expression for the conditional probability \(p(S|\vec {x})\) of observing an *N*-dimensional spike train (spike raster) *S* for a given input signal \(\vec {x}\), which may include external stimulation and/or previous activity of the network itself. Given all the inputs of a neuron, we assume that its probability of spiking is independent of the other neurons (conditional independence). This allows us to factorize \(p(S|\vec {x})={\prod }_{i=1}^{N}p_{i}(S_{i}|\vec {x})\), where \(p_{i}(S_{i}|\vec {x})\) is the probability that the *i*-th neuron, within the recording time [*T* _{0},*T* _{1}], produces a spike train *S* _{ i } conditioned on the input \(\vec {x}\). Therefore, in what follows we focus on the probability \(p_{i}(S_{i}|\vec {x})\) of a single neuron.

**K**

_{ i }. This transformation models synaptic and dendritic filtering, input summation and leaky integration in the soma. The result \(\mathbf {K}_{i}\vec {x}\) is a one-dimensional quantity that is analogous to the membrane potential of a point neuron model. Second, this transformed one-dimensional signal is fed into a nonlinear probabilistic spiking mechanism, which works by sampling from an inhomogeneous Poisson process with an instantaneous rate (conditional intensity function) given by \(\lambda _{i}(t|\vec {x})=f_{i}(\mathbf {K}_{i}\vec {x})\). Here,

*f*

_{ i }(⋅) is a function that captures the nonlinear properties of the neuron. Both the linear filter

**K**

_{ i }and the nonlinearity

*f*

_{ i }are specified by

*𝜃*

_{ i }, a set of parameters that describes the characteristics of the

*i*-th neuron. The schematic of this model is shown in Fig. 1.

*L*(

*𝜃*|

*S*) and expand it as

*X*) is the spike raster

*S*. In the last step of Eq. (1) we have introduced the single neuron log-likelihood \(\mathcal {L}_{i}=\log p_{i}(S_{i}|\vec {x})\).

*λ*

_{ i }(

*t*) produces the spike train

*S*

_{ i }={

*t*

_{ i,k }}, 1≤

*k*≤

*q*

_{ i }, where

*T*

_{0}≤

*t*

_{ i,k }≤

*T*

_{1}and

*q*

_{ i }is the number of spikes of the

*i*-th neuron. This probability is (Brillinger 1988)

*t*

_{ i,0}=

*T*

_{0}. Here, for each spike time

*t*

_{ i,k }, we multiply the (survival) probabilities \(e^{-{\int }_{\!\!t_{i,k-1}}^{t_{i,k}}\lambda _{i}(t)dt}\) of not producing a spike in (

*t*

_{ i,k−1},

*t*

_{ i,k }) with the intensity

*λ*

_{ i }(

*t*

_{ i,k }) at

*t*

_{ i,k }. Finally, we factor in the probability \(e^{-{\int }_{\!\!t_{i,q_{i}}}^{T_{1}}\lambda _{i}(t)dt}\) of not producing a spike in the recording time \((t_{i,q_{i}},T_{1}]\), which remains after the last spike. The function \(L_i (\theta _i | S_i)=p_{i}(S_{i}|\vec {x})\) is known as the point process likelihood (Snyder and Miller 1991).

*k*≤

*q*

_{ i }of the

*i*-th neuron. The first term of this expression rewards high intensity at times

*t*

_{ i,k }when the spikes of the

*i*-th neuron have been emitted, and the second term penalizes high intensity when no spikes have been observed. Different numbers of spikes

*q*

_{ i }render the absolute values of \(\mathcal {L}_{i}\) difficult to compare among different neurons, but play no role when maximizing \(\mathcal {L}_{i}\) with respect to

*𝜃*

_{ i }.

### 2.2 Conditional intensity model for a recurrent neural network

In order to investigate the recurrent aspects of the dynamics of the system, we define the observable input signal \(\vec {x}\) for each neuron as the history of spikes recorded in the network up to a given point in time, including the spikes of the *i*-th neuron itself (which are used to model the refractory properties of the neuron). It is possible to include external inputs in this formulation, however this is not an option that we have pursued in the current work. Below follows a detailed discussion of the different components of the model as presented in Fig. 1.

For simplicity, we assume that the effect of each incoming spike can be modeled as an instantaneous current injection. The spike train *S* _{ j } of the *j*-th neuron as a function of time is expressed as \(s_{j}(t)= {\sum }_{k=1}^{q_{j}} \delta (t-t_{j,k})\), where *t* _{ j,k } is the *k*-th spike of the *j*-th neuron. Each spike then elicits an exponential post-synaptic response in the neuron, due to the filtering properties of the membrane, \(h_{i}(t)=H(t)\exp \left \{ -{t}/{\tau _{i}}\right \} \), where *t* is the time since spike arrival, *τ* _{ i } is the membrane time constant of the neuron, and *H*(*x*)={1 if *x*≥0, else 0} is the Heaviside function, which ensures the causal relationship between the stimulation and the response. Note that while the propagation of spikes is assumed to happen instantaneously in the formulation above, the incorporation of delays will be discussed in detail later in Section 2.4.

*J*

_{ i0}will be used later to set a base level of activity of the unit in the absence of inputs. Differentiation of Eq. (3) yields the first-order ordinary differential equation of the leaky integrator

*U*

_{ i }(

*t*) can be interpreted as the membrane potential of the

*i*-th neuron, while

*J*is the synaptic connectivity matrix and each of its elements

*J*

_{ i j }denotes the strength (synaptic weight) of the connection from the

*j*-th to the

*i*-th neuron. Due to its simplicity, Eq. (4) leads to highly efficient algorithms (discussed in Section 2.3 and 2.6) to evaluate the membrane potential and the conditional intensity function of the neurons, beyond previously reported more general parallelization techniques (Chu et al. 2006). The membrane potential and the intensity are, in turn, needed to compute the values of the likelihood function and its gradient.

In Eqs. (3) and (4), positive and negative values of *J* _{ i j } correspond to excitatory and inhibitory connections respectively, and zero values denote the lack of a connection between two cells. Note that as formulated, this model does not ensure compliance with Dale’s law (according to which each neuron can form synapses of only one type). However, we will show that this is an essentially negligible source of errors in the reconstructions presented below.

*δ*

*u*>0 can be considered as the inverse “gain” of the nonlinearity. In the derivations that follow we will assume

*δ*

*u*=1 in order to simplify the expressions without loss of the generality, as different gains can be accommodated by rescaling the synapse weights

*J*

_{ i j }and the baseline potential

*J*

_{ i0}accordingly. In the absence of input spikes,

*U*

_{ i }=

*J*

_{ i0}, which leads to the base rate

The model as formulated above is similar to the widely used cascade LNP model (Simoncelli et al. 2004), but in addition to the activity of the other cells in the ensemble, it also incorporates the spiking history of the neuron itself through its self-connection *J* _{ i i }. An intuitive biological interpretation of this class of models, also known as the spike-response model with escape noise, in relation to the conventional integrate-and-fire model is given in Brillinger (1988) and Gerstner et al. (2014). Here, in contrast to the approaches taken in previous studies (Song et al. 2013; Citi et al. 2014; Ramirez and Paninski 2014), we drastically simplify both the conditional intensity model for a single neuron and the interaction kernels. This makes the numerics in our method amenable to a highly efficient implementation as discussed in Section 2.6.

Given that *f* _{ i }(⋅) is both a convex and log-concave function of \(U_{i}=\mathbf {K}_{i}\vec {x}\), and the space of possible {**K** _{ i }} is convex, it can be shown that the log-likelihood function of such problems is concave and does not have any non-global local extrema (Paninski 2004). Thus the log-likelihood function \(\mathcal {L}_{i}\) of the model as formulated above is concave in \(\theta _{i}\subset \{J_{ij}\}_{0\leq j\leq N}\) (note, however, that *τ* _{ i } is not included in *𝜃* _{ i }; the recovery of the time constants will be addressed separately). A proof of the concavity of \(\mathcal {L}_{i}\) for our specific choice of kernels and link function is given in Appendix A. Since the sum of concave functions is again concave, the full log-likelihood \(\mathcal {L}={\sum }_{i=1}^{N}\mathcal {L}_{i}\) is concave as well. Consequently, there exists a unique set of parameters *𝜃* that characterize the network model that is most likely to exhibit a given recorded activity. These parameters *𝜃* can be efficiently identified via gradient ascent based nonlinear optimization methods applied to \(\mathcal {L}\). Moreover, due to the separability of \(\mathcal {L}\)(1), in order to recover *𝜃*={*𝜃* _{ i }}, one can maximize the individual log-likelihood functions \(\mathcal {L}_{i}\) for each recorded unit, instead of maximizing the complete log-likelihood function \(\mathcal {L}\).

Since the experimental techniques to obtain simultaneous recordings of thousands of units are becoming increasingly accessible, in this work we are targeting *N*≥1000. However, even if the number of variables is reduced from the \(\mathcal {\sim O}(N^{2}=10^{6})\) required for the complete log-likelihood function to the \(\mathcal {\sim O}(N=10^{3})\) required for the log-likelihood function of an individual neuron, this is still a high-dimensional convex optimization problem. It can only be solved in practice using gradient based methods, for which the analytical closed form expressions for the log-likelihood function and its gradient are both available, and amenable to efficient evaluation. In the following we derive these expressions for the postulated model.

### 2.3 Closed form expressions

*U*

_{ i }(

*t*), computing \(\mathcal {L}_{i}^{\Sigma }\) is a matter of a simple algebraic substitution, while the efficiency of computing \(\mathcal {L}_{i}^{\int }\) depends on whether it is possible to find this primitive analytically.

#### 2.3.1 Recurrence formula for the membrane potential

**K**

_{ i }(exponential post-synaptic potential plus baseline potential) allows us to obtain the required closed form for

*U*

_{ i }(

*t*) because it obeys the leaky integrator dynamics (4). The solution of Eq. (4) from

*t*

_{ k }to

*t*in the absence of input spikes

*s*

_{ j }(

*t*) is \(U_{i}(t)=(U_{i}(t_{k})-J_{i0})\exp \left \{ -\frac {t-t_{k}}{\tau _{i}}\right \} +J_{i0}\). This expression is valid at any time

*t*between two consecutive observed spikes

*t*

_{ k },

*t*

_{ k+1}∈

*S*, where

*S*={

*t*

_{ k }} is the (ordered) set of all recorded spikes of the network. At the borders of each of those intervals, the value of

*U*

_{ i }(

*t*

_{ k+1}) is increased by the contribution of the corresponding incoming spike:

*j*refers to the neuron that emitted a spike at time

*t*

_{ k+1}; if spikes from multiple neurons

*j*

^{1,2,3,…}arrive at time

*t*

_{ k+1}, the contributions \(J_{ij^{1,2,3,\ldots }}\) have to be added. We will refer to Eq. (8) as the key recurrence formula in the following.

The formula (8) for *U* _{ i }(*t* _{ k+1}) makes it possible to find the value of the membrane potential of the neuron at the spike time *t* _{ k+1} given the previous value at time *t* _{ k } by computing only one exponential function. It is substantially more efficient in terms of computation than naively summing up the contributions from all spikes that happened at *t*<*t* _{ k } for each point in time *t* _{ k }. In particular, for kernels with infinite memory like the exponential kernels *h* _{ i }(*t*) employed here, the recurrence formula (8) is crucial to avoid an explosion of the computational costs when evaluating the log-likelihood on large datasets in continuous time.

#### 2.3.2 Evaluating the likelihood

*t*

_{ k }to

*t*

_{ k+1}:

*t*

_{0}=

*T*

_{0}and

*t*

_{ q+1}=

*T*

_{1}are the start and end of the recording. The integral contained here has a known closed form, so

*x*) is a special function (exponential integral) defined as \(\text {Ei}(x)=-{\int }_{\!\!-x}^{\infty }\frac {e^{-t}}{t}dt\) for real nonzero values of

*x*. For a proof of the equivalence of Eqs. (9) and (10) see Appendix B; the numerical computation of this function is discussed below in Section 2.6.1. The summands of Eq. (10) are independent, and therefore the evaluation of \(\mathcal {L}_{i}^{\int }\) lends itself to trivial parallelization.

#### 2.3.3 Evaluating the gradient

*N*, since \(\mathcal {L}_{i}\) has \(\mathcal {O}(N)\) partial derivatives that all need to be evaluated at each step of the optimization. The parameters of \(\mathcal {L}_{i}\) are \(\theta _{i}=(J_{i0},{\dots } J_{iN})\). For convenience, let us first introduce the terms

*j*≥1, can be interpreted as the putative response of the

*i*-th neuron to the input spikes from the

*j*-th neuron, that is going to be scaled by

*J*

_{ i j }, cf. (3). The derivatives of \(\mathcal {L}_{i}\)(7) with respect to

*J*

_{ i j }can then be expressed as

*q*

_{ i }is the number of spikes of the

*i*-th neuron, and {

*t*

_{ i,k }}=

*S*

_{ i }are the points in time when the

*i*-th neuron emitted a spike. For

*j*=0, Eq. (12) becomes \(\frac {\partial }{\partial J_{i0}}\mathcal {L}_{i}=q_{i}-\mathcal {L}_{i}^{\int }\). This means that at a maximum of \(\mathcal {L}_{i}\), the baseline potential

*J*

_{ i0}(and so the base rate

*c*

_{ i }(6)) is set such that the number of spikes

*q*

_{ i }equals the expected total number of spikes of the GLM, \(\mathcal {L}_{i}^{\int }\). Further, in order to evaluate (12) for the cases when

*j*≥1, we have defined the symbols \(\partial _{ij}^{\Sigma }\) and \(\partial _{ij}^{\int }\) analogous to Eq. (7).

*ν*

_{ i j }(

*t*

_{ i,k }) for

*j*≥1 can be obtained using a recurrence formula just like for the membrane potential

*U*

_{ i }(

*t*

_{ k })(8); in fact, \(\nu _{ij}(t)=\frac {\partial }{\partial J_{ij}}U_{i}(t)\), cf. (11). Hence,

*ν*

_{ i j }(

*t*) obeys leaky integrator dynamics like

*U*

_{ i }(

*t*), which can be obtained by differentiating Eq. (4) by

*J*

_{ i j }and reinserting Eq. (11). Accordingly,

*ν*

_{ i j }(

*t*) decays exponentially in between spikes \(\nu _{ij}(t_{j,k}<t<t_{j,k+1})=\nu _{ij}(t_{k})\exp \left \{ -\frac {t-t_{j,k}}{\tau _{i}}\right \} \), and we find the recurrence formula

*ν*

_{ i j }(

*t*

_{ i,k }) then yields \(\partial _{ij}^{\Sigma }\).

It is also important to mention that *ν* _{ i j }(*t*)(11) and, consequently, \(\partial _{ij}^{\Sigma }\) in Eq. (12) do not depend on parameters *𝜃* _{ i } and therefore need only be computed once at the beginning of the optimization. However, even though we can use the formula \(U_{i}(t)={\sum }_{j=0}^{N}J_{ij}\nu _{ij}(t)\), for large *N* it is more expensive to compute *U* _{ i }(*t*) by summing up weighted contributions of *ν* _{ i j }(*t*) than by using Eq. (8) as explained above.

*U*

_{ i }(

*t*)(8) and

*ν*

_{ i j }(

*t*)(13), the closed form of \(\partial _{ij}^{\int }\) in Eq. (12) can be expressed as follows:

*t*

_{0}=

*T*

_{0}and

*t*

_{ q+1}=

*T*

_{1}are the start and end of the recording. Unlike \(\partial _{ij}^{\Sigma }\), this expression needs to be re-evaluated at every optimization step, but as with Eq. (10), the elements of the sum are independent from each other and can therefore also be efficiently parallelized.

### 2.4 Handling transmission delays

In the discussion above, the communication of spikes between the neurons was implicitly assumed to happen instantaneously. Of course, in reality spikes incur transmission delays, which strongly affects the dynamics of the network.

Fortunately, the effects of combined synaptic and axonal delays can be easily incorporated into the described model: thanks to the separability property, we can optimize the parameters for each neuron independently, and feed every optimization for different neurons with its own modified dataset, containing the incoming spike times from other neurons arriving as the target neuron actually received them.

*D*, it is only necessary to shift each spike train

*S*

_{ j }in the recorded raster

*S*by the corresponding delay at the beginning of the optimization for the

*i*-th neuron, such that the membrane potential of this neuron is affected at the point in time when the incoming spikes from the

*j*-th neuron have reached their target, and not immediately as they were fired (and recorded):

The transformation above has to be applied with one exception: the elements of the sum in \(\mathcal {L}_{i}^{\Sigma }\) (and, accordingly, \(\partial _{ij}^{\Sigma }\)) have to be evaluated at time points *S* _{ i } when the *i*-th neuron actually produced a spike, and not at time points \(\widehat {S}{}_{i}=S_{i}+D_{ii}\), when this spike has reached the neuron through the “self-connection” and provoked a depression of its membrane potential, which models the refractory properties of the neuron.

In other words, in order to correctly evaluate the expressions Eqs. (7) and (12) while taking into account transmission delays, one must compute the values of \(\widehat {U}_{i}(t)\) and \(\widehat {\nu }{}_{ij}(t)\) using the *modified* raster \(\widehat {S}\), but at time points *S* _{ i } of the *original* raster *S*, and substitute these values in the elements of the sums \(\mathcal {L}_{i}^{\Sigma }\) and \(\partial _{ij}^{\Sigma }\) respectively, instead of summing up the elements taken at times \(\widehat {S}_{i}\). In the following, we omit the “hats” for notational convenience.

### 2.5 Regularization of the model

Substantial improvements in the quality of the network reconstruction can be achieved if the model presented above is subjected to standard regularization techniques. These techniques enhance the accuracy of the inference procedure by integrating additional prior knowledge about the system into the optimization process (Meinshausen and Bühlmann 2006; Ravikumar et al. 2010). For instance, we can impose box constraints on reasonable values of the synaptic connection matrix *J* _{ i j } or base rates *c* _{ i }, and complement this with a choice of more sophisticated methods, such as *ℓ* _{1} or *ℓ* _{2} regularization, exploiting assumed sparsity or smoothness of the expected result, respectively (Chen et al. 2011).

*ℓ*

_{1}regularization (Tibshirani 1996) has a straightforward Bayesian interpretation in our setting: by penalizing the log-likelihood function (2) with the sum of the absolute values of the synaptic weights

*J*

_{ i j }, we impose a sparsity-inducing Laplace prior on the sought-for solution, thereby performing a

*maximum a posteriori*(MAP) estimation. Here the strength of the penalty

*α*reflects the firmness of our belief in the sparseness of the network connectivity:

Possible overfitting due to an inadequate choice of the regularization parameter *α* can be prevented by separating the dataset into two parts to cross-validate the recovered synaptic weights, and, in the case that the available data is too scarce, more elaborate techniques such as *K*-fold cross-validation and other cross-validation types (Kohavi 1995) can be employed.

### 2.6 Practical implementation

The mathematical components described above make it possible to reproduce our estimation procedure. However, we found that without employing additional numerical methods, a naive implementation would be way too slow for practical use. In the following we outline the techniques that helped us to boost the optimization speed by many orders of magnitude, bringing the computational requirements to perform estimations of the connectivity for the networks of \(N\sim \mathcal {O}(10^{3})\) neurons into a practical range for plausible amounts of experimental data.

#### 2.6.1 Efficient evaluation

From the computational perspective, a program that performs the parameter estimation would typically consist of a nonlinear optimization routine, which is provided with callback procedures that are repeatedly called in order to evaluate the objective function (2) and its gradient (12) for any given set of parameters. Hence, the cornerstone guiding principle to achieve best performance is to carefully consider the CPU time versus memory consumption trade-offs, and cache as many values for these callbacks as feasible.

As the values of *U* _{ i }(*t*) for *S*={*t* _{ k }} (all spikes of the network) are needed in order to evaluate both the log-likelihood function and its gradient, it makes sense to pre-compute these values at the beginning of the optimization step. Additionally, as previously noted, the values of *ν* _{ i j }(*t*) do not depend on the parameters *𝜃*, and therefore both *ν* _{ i j }(*t*) and \(\partial _{ij}^{\Sigma }\) can be pre-computed during the first optimization step, and re-used in all subsequent steps. Likewise, it is important to consider the costs of calculating transcendental functions; whereas they might seem negligible at the first sight, the time taken to compute some 10^{10} exponentials every step is considerable. Therefore, pre-computing the values of all sub-expressions that do not depend on the parameters, and, in particular, \(\xi _{t_{k}}=\exp \left \{- (t_{k+1}/-t_{k}){\tau _{i}}\right \} \) is another possibility to save large amounts of CPU time.

In any case, we recommend using iterative profiling in order to select the relevant optimization targets to add each next level of caching, since, as a general rule, the more caches there are, the more complicated and error-prone it is to keep them consistent and up to date with respect to the changes in parameters. Additionally, this avoids the situations when a sizeable amount of work is invested only to gain minor improvements in speed, due to runtime actually being dominated by different code paths than anticipated.

We observed that the optimization algorithms are (unsurprisingly) sensitive to the precision of the evaluation of the objective function and its gradient, and especially to the consistency between the two. Therefore we rejected using numerical approximations to the gradient, such as values computed using the central differences formula, and employed analytically derived expressions instead. We have also found that better precision of the objective function leads to faster convergence. This particularly concerns the accurate approximation of the exponential integral in Eq. (10). In general, finding an efficient method to evaluate Ei(*x*), which is a crucial part of Eq. (10), poses a significant computational challenge. However, high-quality rational approximations exist in the literature (Cody and Thacher 1969), which make it as fast as evaluating low-order polynomials. In our implementation, we rely on the approximations devised by John Maddock using a custom Remez code, which are part of the Boost C++ library.^{1} These approximations are not only highly accurate, but also the fastest that are available to us.

#### 2.6.2 Parallelization and distribution

As the sweeping growth of the clock speeds in the last couple of decades seems to have saturated, the focus is increasingly shifting towards increasing parallelism, and nowadays multicore CPUs are a *de facto* standard, rather than rare marvels. Therefore, suitability for parallelization is becoming a critical feature to discriminate the algorithms that are appropriate for large-scale data analysis. In this section we discuss the parallelization strategies applicable to the model described above.

Owing to the separability of the problem, the highest level approach to parallelize the execution of the optimization is to launch several estimations for different neurons in parallel. This results in a perfect scaling for *N* _{t}≤*N*, where *N* _{t} is the number of simultaneously executed hardware threads. This is clearly a very attractive option due to the relative simplicity of implementation, however, its practical applicability is limited by the amount of the available memory per thread, which quickly becomes a bottleneck for larger networks and bigger amounts of data.

A slightly lower-level method is to identify independent elements in the formulae that need to be evaluated at every step of the optimization, and divide this work among several threads within one running process. The summands of \(\mathcal {L}_{i}^{\Sigma }\), \(\mathcal {L}_{i}^{\int }\), \(\partial _{ij}^{\Sigma }\) and \(\partial _{ij}^{\int }\) as defined in Eqs. (7), (10), (12) and (14) are all amenable to that kind of processing. This approach is advantageous to utilize all usable threads from within one process, but its scalability is limited by both the amount of the available memory on a single compute node (as above), and the serial part of the computations, which cannot be parallelized. In our model, it is mainly the calculation of the membrane potential *U* _{ i }(*t*)(8) and the membrane responses *ν* _{ i j }(*t*), because each value in the recurrence formulae depends on the previous one. The membrane responses *ν* _{ i j }(*t*) are less of a problem, since they can be pre-computed at the beginning of the optimization as explained above, if one is willing to trade memory consumption for performance. Alternatively, *ν* _{ i j }(*t*) can be computed in parallel, which can be faster than fetching the results from memory for a very high number of threads and low memory bandwidth.

We have also explored the possibility of distributing the estimation across several compute nodes, which is not only necessary in order to utilize larger numbers of threads than available on one node, but also allows the computation to make use of the additional memory when the problem gets too large to fit into one machine’s RAM. The most straightforward distribution scheme is to designate one process (rank) to perform serial computations required for every optimization step, broadcast the results and parameters to other ranks, have them do their share of the computations, and, finally, collect the results. The biggest advantage of this scheme lies in its ease of implementation: the communication pattern is very clear, and the code can largely remain unchanged except for the need of a few additional functions to distribute and collect the data.

In our implementation, we performed the calculation of the membrane potential *U* _{ i }(*t*), the log-likelihood function \(\mathcal {L}_{i}\) and \( \partial \mathcal {L}_{i}/{\partial J_{i0}}\) on Rank 0, and evenly divided the work to compute \( \partial \mathcal {L}_{i}/{\partial J_{ij}}\), *j*≥1, among all other ranks. This system scales (almost) linearly up to the point when the amount of time needed to perform the computations on Rank 0 exceeds the amount of time it takes to compute the gradient distributed to all other ranks. Since it takes several orders of magnitude more time to calculate \(\mathcal {L}_{i}\) than \( \partial \mathcal {L}_{i}/{\partial J_{ij}}\), we have found that for *N*=1000 we can easily distribute each single task up to *N* _{r}=10…20 ranks.

For production estimations, we combined all three approaches outlined above. The highest level of parallelization was left up to the batch system: for each estimation, we generated and submitted the job scripts for every neuron and let the scheduler optimally backfill the queue. The code was run with *N* _{t}=8…16, depending on the amount of hardware threads available per processor, and *N* _{r}=10…20, depending on the amount of available memory per processor and the requirements of the particular estimations. For estimations of size *N*=1000, this hybrid approach allowed us to scale almost linearly up to \(\mathcal {O}(N_{\mathrm {t}}\times N_{\mathrm {r}}\times N=10^{5})\) cores.

In this context, it becomes clear why not only the convexity, but also the separability property of the optimization problem discussed in Section 2.2 is crucial to our model. In a typical estimation, as described in Section 3, 1 hour recording of *N*=1000 neurons spiking at ∼5 s^{−1} would contain ∼10^{7} spikes, so the intermediary data to be held in RAM during the optimization would need around ∼10^{14}=10×10^{7}×(10^{3})^{2} bytes or 100 TB of storage capacity. This calculation assumes that the main contribution comes from the pre-computed matrix of *ν* _{ i j }(*t*) vectors of length 10^{7} stored as doubles and disregards all other factors. From our experience, for some *N* _{r}×*N* _{t}=10^{5} threads at ∼2 GHz the optimization would take an order of magnitude of 30 minutes of walltime to converge after about a hundred of iterations.

Currently, these requirements can be barely satisfied by booking a complete supercomputer such as JUROPA,^{2} and any substantial increase in the number of units, or in the amount of data to be processed will put the problem beyond our reach. However, while the number of parameters of the complete log-likelihood function \(\mathcal {L}\) in our formulation is \(\mathcal {O}(\theta )\sim N^{2}\), thanks to the above-mentioned separability property, the number of parameters of \(\mathcal {L}_{i}\) is linear in the number of units, \(\mathcal {O}(\theta _{i})\sim N\). Not only does this present major practical advantages such as easier scheduling of smaller jobs, but it also makes it possible to solve larger problems at all by proportionally trading the execution time for the amount of resources allocated to the optimization process.

#### 2.6.3 Technical realization

Our model was implemented in Python, an increasingly popular language in the field of computational neuroscience. It relies upon the NumPy and SciPy scientific libraries^{3} for essential data structures and algorithms. We used Cython^{4} in order to bind to the OpenMP-parallelized computational kernels, that we extracted and re-wrote in C++ for performance reasons, and in order to access the mathematical functions from Boost C++ library. The distribution was implemented using the Python bindings to MPI, mpi4py.^{5}

The optimization was performed via the NLopt^{6} package by Steven G. Johnson using the low-storage Broyden-Fletcher-Goldfarb-Shanno method (Liu and Nocedal 1989) with support for bound constraints (Byrd et al. 1995) implemented by Ladislav Luksan (L-BFGS-B). We chose to use BFGS instead of the nonlinear conjugate gradient (CG) algorithm, because the former approximates the inverse Hessian matrix of the problem and uses it to steer the search in the parameter space. This results in improved convergence at the cost of higher iteration overhead. Since in our case the computation of the objective function is substantially more expensive, this trade-off is worthwhile.

As a stopping condition, we used a criterion based on the fractional tolerance of the objective function value. The optimization was terminated if \(\eta =|{\Delta }\mathcal {L}|/|\mathcal {L}|\), where \({\Delta }\mathcal {L}\) is the decrease in the function value from one iteration to next, reached the threshold of \(\tilde {\eta }\). The value of \(\tilde {\eta }\) was selected close to the machine epsilon for the double precision floating point type, as requesting even lower tolerance would not yield a more accurate solution; the typical choice was \(\tilde {\eta }\leq 10^{-15}\).

*ℓ*

_{1}regularized optimizations, it turned out that all gradient-based algorithms we tried were very much affected by the non-smoothness at zero, introduced by the regularization term in Eq. (16). A thorough review of the existing approaches to address this issue is presented in (Schmidt et al. 2009); we opted for implementing a smooth

*𝜖*–

*ℓ*

_{1}approximation, originally suggested in Lee et al. (2006):

*J*

_{ i j }(12) have to be adjusted by addition of \(-\alpha J_{ij}/\sqrt {J_{ij}^{2}+\epsilon }\) respectively. We found that this approximation works well in practice for sufficiently small values of

*𝜖*<10

^{−7}and enables us to use the L-BFGS-B algorithm without modifications. Additionally, we imposed bound constraints on the model parameters as discussed in Section 2.5; typical constraint ranges were |

*J*

_{ i j }|<50 mV for synaptic weights and \(0.001\text { s}^{-1}<c_{i}<100\text { s}^{-1}\) for base rates. The recordings were truncated to the first and last recorded spikes,

*T*

_{0}=

*t*

_{1}and

*T*

_{1}=

*t*

_{ q }, where

*q*is the total number of recorded spikes.

## 3 Results

Glossary of abbreviations

EM | expectation-maximization |

GLM | generalized linear model |

GMM | Gaussian mixture model |

KDE | kernel density estimation |

LIF | leaky integrate-and-fire |

MER | misclassification error rate |

MLE | maximum likelihood estimation |

| probability density function |

### 3.1 Random balanced network of GLM neurons

As an initial testbed for our method, we selected a random balanced neural network of excitatory and inhibitory neurons in the asynchronous irregular (AI) spiking regime (Brunel 2000). Random networks do not have any particular structural features that can be exploited by the optimizer in order to improve the quality of the reconstruction, and hence in this sense they represent a “worst-case” type of input that is particularly useful for benchmarking purposes. Such networks are commonly studied using the leaky integrate-and-fire (LIF) neuron model. However, in order to be able to interpret the follow-up experiments, we first chose to assess the performance of our estimation method under idealized conditions, in which the simulated and estimated neuron and synapse models coincide: the GLM neuron model as described in Section 2 and simple synapses with exponential post-synaptic potentials.

As discussed in Section 2.2, given several conditions that our GLM satisfies, there is a unique maximum likelihood parameter set for the estimated network model (Paninski 2004). In the limit of an infinite amount of spike data used for model estimation and arbitrarily precise calculations, our method is thus bound to recover the true parameters of the simulated model. Hence, testing the method under idealized conditions, but for finite datasets, allows us to distinguish errors that are purely due to the limited length of the observations and restricted machine precision, from those due to a mismatch between the dynamics of the neuron and synapse models used to generate the data, and the dynamics of the models used to reconstruct the network.

The test network consisted of *N*=1000 GLM neurons with 80 % : 20 % proportion of excitatory to inhibitory neurons (“pp_psc_delta” model in NEST nomenclature, with a base rate *c*=5 s^{−1}, membrane time constant of *τ*=20 ms and a resting potential of *V* _{r}=0 mV). The nonlinearity gain of the neurons was set to *δ* *u*=4 mV as in Jolivet et al. (2006), which defines the scaling and units of a single post-synaptic potential via Eq. (5) (*δ* *u*=1 as assumed previously in Section 2 for the sake of convenience would make it unitless). Each connection was realized independently with a connection probability of *𝜖*=0.2 (Erdős-Rényi p-graph). The neurons were connected by synapses with exponential post-synaptic potentials with a peak amplitude of *J* _{e}=1 mV for excitatory and *J* _{i}=−5 mV for inhibitory synapses, and a transmission delay of *d*=1.5 ms. A strong inhibitory self-connection with *J* _{s}=−25 mV and a transmission delay of *d* _{s}=Δ*t* was used to model post-spike effects. The simulation progressed in time steps of Δ*t*=0.1 ms (resolution) and the simulation time was *T*=1 hour. The average firing rate of the neurons was *ν*=4.2 s^{−1}. The recorded spike trains were fed to the estimation method, assuming known values of the time constant *τ*, the transmission delays *d* and the delay of the self-connection *d* _{s}. The method produced estimates of the synaptic weight matrix *J* _{ i j } and the base rates {*c* _{ i }} for all neurons. The original and reconstructed synaptic weight matrix for this experiment are presented in Fig. 2. Throughout this text we refer to {*J* _{ i j }}_{1≤i,j≤N, i≠j } as the weight matrix; the self-connections {*J* _{ i i }}_{1≤i≤N } and the baseline potentials {*J* _{ i0}}_{1≤i≤N } are treated separately.

In order to evaluate the quality of the reconstruction, we analyzed the resulting distributions of recovered synaptic weights and base rates, as shown in Fig. 3. Whereas the probability density function (PDF) of the original distribution of synaptic weights can be described as a sum of three *δ*-functions (for excitatory, inhibitory and null connections respectively), the peaks in the reconstructed distribution are broader due to the finite duration of the recording and limited machine precision, to the extent that for realistic values of parameters, there is a degree of overlap between the components of the distributions that represent excitatory and null connections. We noted that the amplitude of the noise that causes the broadening decreases approximately in inverse proportion to the square root of the duration of the recording (data not shown), however, we selected *T*=1 hour as a reasonable standard amount of input data to mimic conditions where the duration of the recording is limited due to experimental and computational constraints.

This circumstance thus makes it difficult to identify weak excitatory connections unambiguously, and therefore an advanced approach to classification was needed to obtain optimal network reconstruction. To this end, we fitted a Gaussian mixture model (GMM) with a fixed number of components (*n*=3) to the reconstructed synaptic weights, assuming that synaptic connections, in general, can be either excitatory or inhibitory, or absent. We used an expectation-maximization (EM) algorithm to obtain a maximum likelihood estimate (MLE) of the GMM parameters (mixing weights, means and variances of the individual components), and classified the synaptic weights accordingly. The fitting and classification was performed using a Python implementation of GMM (sklearn.mixture.GMM) provided by the scikit-learn toolkit (Pedregosa et al. 2011). In order to reconstruct the PDFs of the base rates and self-connections, we used both the histogram function from the NumPy library and the Gaussian kernel density estimation (KDE) code from the SciPy library.

*%*.

Breakdown of classification errors for the GLM random network

Connection type | Errors | FP | FN | ND |
---|---|---|---|---|

Excitatory | 7 300 | 62 % | 38 % | 15 % |

Inhibitory | 0 | |||

Unconnected | 7 300 | 38 % | 62 % | — |

Total errors | 0.73 % |

### 3.2 Random balanced network of LIF neurons

Having established the baseline performance in ideal conditions, we designed our next experiment to gauge the influence of mismatch between the neuron and synapse models used to generate the data and those used to reconstruct the network. To this end, we generated data with the commonly used, more complex and realistic LIF neuron model with *α*-shaped post-synaptic currents (PSCs). We then carried out the reconstruction as before assuming our simplified GLM neuron model and synapses with exponential post-synaptic potentials. Another important point is that whereas in the previous experiment we assumed that the membrane time constant *τ* and transmission delays between the neurons *d* are known in advance, this is certainly not the case in the laboratory setting, and hence a principled way of estimating these parameters is required in order to analyze real physiological data.

To generate the test data, we wired a network similar to the one described in the previous section, but using a LIF instead of a GLM neuron model. As before, we used *N*=1000 neurons with 80 % : 20 % ratio of excitatory to inhibitory cells, connection probability of *𝜖*=0.2 (each connection was realized independently), transmission delay of *d*=1.5 ms, simulation resolution of Δ*t*=0.1 ms. Synaptic weights were set to \(\hat {J}_{\mathrm {e/i}}=J_{\mathrm {e/i}}\times w\), with *J* _{e}=1 mV and *J* _{i}=−5 mV. The latter (*J* _{e} and *J* _{i}) were again interpreted as peak PSP amplitudes, where *w*=*w*(*τ* _{m},*τ* _{s},*C*) was the scaling factor (specific to the post-synaptic neuron) selected such that an incoming spike passing through a connection with the synaptic weight of *w* would evoke a PSP with the maximum amplitude of 1 mV. The parameters of the LIF model (“iaf_psc_alpha” in NEST nomenclature) were chosen as follows: membrane capacitance *C*=250 pF, membrane time constant *τ* _{m}=20 ms, synaptic time constant *τ* _{s}=0.5 ms, refractory time *t* _{r}=2 ms, firing threshold *𝜃*=20 mV, resting potential *V* _{r}=0 mV and reset to *V* _{r} after each spike. This time, additional to the synaptic input from other simulated neurons, each neuron received independent Poisson process excitatory inputs at a rate of \(\nu _{\mathrm {e}}=1779\text { s}^{-1}\) and inhibitory inputs at \(\nu _{\mathrm {i}}=0.2\times \nu _{\mathrm {e}}=356\text { s}^{-1}\). These external inputs represent the influence of neurons that are not part of the simulation, and are necessary to achieve asynchronous and irregular activity as in cortical networks (Brunel 2000). The simulation time was set to *T*=2 hours and the data was cut into training and validation parts of *T* _{t}=*T* _{v}=1 hour as explained below. The average neuron firing rate was *ν*=4.2 s^{−1}, and so matched the average neuron firing rate of the network of the GLM neurons presented above.

*τ*and

*d*for this experiment, we applied a cross-validation procedure. It is important to note that we are not expecting to obtain exactly

*τ*=

*τ*

_{m}=20 ms and

*d*=1.5 ms due to mismatch between the LIF with

*α*-shaped PSCs and GLM with exponential PSPs models. Instead, we want to recover the optimal parameters

*τ*and

*d*for the GLM model to produce most similar dynamics to the recorded spike trains from the LIF model. We split the available data into a training and a validation dataset, and performed reconstructions for a subset of

*N*

_{s}=75 neurons on the training dataset varying one parameter, while keeping the other one fixed. The resulting parameter estimates

*𝜃*

_{ i }were then used to calculate the log-likelihood function \(\mathcal {L}_{i}\) on the validation dataset. Two different datasets (training and validation) were used in order to ensure that the chosen values of the parameters generalize, and are not specific to the training sample. The validation curves are shown in Fig. 4a, c (the curves for the training dataset look identical); note that they all have an easily identifiable maximum. Subsequently, we averaged the locations of the maxima for all trials and performed another cross-validation run (Fig. 4d, b) for updated values of the parameters. Repeating this procedure of alternatively fixing one parameter and performing cross-validation for another one would lead us to a local extremum in the (

*τ*,

*d*) parameter space. However, we opted to stop after only a few iterations because the procedure is computationally expensive, and in order to asses if a sub-optimal choice of

*τ*=10 ms and

*d*=1.7 ms would lead to acceptable estimation results.

*τ*=10 ms and

*d*=1.7 ms through the cross-validation procedure, we used these values to estimate the connectivity and base rates. The results of the connectivity reconstruction on the training dataset were processed in the same way as in the previous subsection and are presented in Fig. 5, with further details on the classification of synaptic connections in Table 3. We find that the reconstruction quality as defined by classification into the groups of excitatory, inhibitory and null connections closely matches the performance on the ideal dataset analyzed in the previous section, despite the mismatch in models and the suboptimal choice of

*τ*and

*d*. Note that in this experiment, the recovered values of synaptic weights in mV cannot be compared directly to the ones that were used in the simulation which produced the data due to the differences between GLM and LIF models, unlike in the first experiment described in Section 3.1. However, this does not matter for the purposes of classification.

Breakdown of classification errors for the LIF random network with *α*-shaped PSCs

Connection type | Errors | FP | FN | ND |
---|---|---|---|---|

Excitatory | 7 020 | 58 % | 42 % | 10 % |

Inhibitory | 2 | 100 % | 0 % | 100 % |

Unconnected | 7 022 | 42 % | 58 % | — |

Total errors | 0.70 % |

### 3.3 Random balanced network with distributed parameters

*J*

_{e},

*J*

_{i},

*d*,

*τ*

_{m}and

*τ*

_{s}for every neuron and synaptic connection, sampled from uniform distributions around each respective mean value (Table 4), which are the same as in the previous experiment. However, instead of trying to recover the individual values of

*τ*

_{ i }for each neuron and

*d*

_{ i }for every connection, we decided to investigate whether it would be still possible to make a useful reconstruction assuming identical “mean” values of

*τ*for all neurons and

*d*for all connections. Additional motivation for this choice is in that cross-validation is a computationally expensive procedure: whereas individual estimation might converge in a matter of minutes, the amount of resources needed to scan a multidimensional parameter grid grows quickly and becomes unmanageable. Therefore, we performed cross-validation on a subset of neurons as described in the previous subsection, and settled for

*τ*=10 ms and

*d*=1.7 ms again (data not shown).

Distribution of parameter values of a random balanced network of LIF neurons

Parameter | Symbol | Range | Spread |
---|---|---|---|

Excitatory weight | | 0.8…1.2 mV | ±20 |

Inhibitory weight | | −4…−6 mV | ±20 |

Transmission delay | | 1…2 ms | ±33 |

Membrane time constant | | 15…25 ms | ±25 |

Synaptic time constant | | 0.3…0.7 ms | ±40 |

*τ*

_{ i }and

*d*

_{ i }, we used rather arbitrarily chosen fixed values for all neurons and connections. However, more importantly, as the components of the original PDF of synaptic weights were broad distributions rather than

*δ*-functions, the resulting recovered distribution components are strongly non-Gaussian. Therefore, in this case the EM procedure for GMM fails to converge to reasonable means and variances, and is no longer a viable choice to perform the classification of connections.

Breakdown of the classification errors for the unregularized and *ℓ* _{1}-regularized (*α*=10) connectivity estimations of a LIF random network with distributed parameters

Unregularized | Regularized, | |||||||
---|---|---|---|---|---|---|---|---|

Connection type | Errors | FP | FN | ND | Errors | FP | FN | ND |

Excitatory | 57 827 | 84 % | 16 % | 17 % | 33 706 | 16 % | 84 % | 3 % |

Inhibitory | 2 226 | 95 % | 5 % | 79 % | 2 673 | 0 % | 100 % | 0 % |

Unconnected | 59 749 | 16 % | 84 % | — | 36 379 | 85 % | 15 % | — |

Total errors | 5.99 % | 3.64 % |

However, instead of engaging in more elaborate statistical modeling to overcome this difficulty, we can take a step back and resort to an unsupervised learning technique called *k*-means clustering (which is actually a simplification of GMM). This method rejects the probabilistic assignment of data points to components, and instead makes the assumption that each point belongs to one (and only one) cluster, to the centroid of which it is closest in terms of Euclidean distance. This simplification leads to sub-optimal classification when the underlying distributions violate these constraints, but the resulting algorithm is fast and robust.

The Voronoi diagrams for *k*-means classification are represented in Fig. 6 as solid lines: the colors show which of the three centroids is closest, in blue, green and red for inhibitory, null and excitatory connections, respectively. By comparing the solid curves and envelope of the colored bars it can be seen that in this case there is a significant overlap between the components contributed by null connections and excitatory connections. Therefore, even the most advanced classification strategies will lead to a substantially higher amount of classification errors than in the previous experiments. The classification data using *k*-means is given in Table 5 (left part).

Nevertheless, the situation can still be considerably improved: here, we exploited the sparsity of the synaptic connection matrix by regularizing the GLM estimation with a *ℓ* _{1} penalty term as explained in Section 2.5. Imposing such a prior on the estimation causes shrinking of the distribution of null connections (Tibshirani 1996) and thus enables better separation between the components. However, the choice of the penalty scaling constant *α* is arbitrary and so we again availed ourselves of a cross-validation procedure to determine the optimal value for our dataset.

*α*on the training dataset are shown in the left panel of Fig. 7. The right panel depicts the subsequent evaluation of the log-likelihood function on the validation dataset. It is important to note that, for optimal results, this procedure should generally be performed for all neurons, and an individual regularization coefficient should be selected for each of the cells. Instead, in order to save computational resources, we only performed it for a subpopulation of neurons and subsequently selected the same value of

*α*=10 for all cells, which is slightly lower than the average, to prevent excessive connection pruning in neurons with small optimal

*α*.

We performed a full *ℓ* _{1}-regularized GLM estimation using *α*=10, still fixing the parameters to *τ*=10 ms and *d*=1.7 ms, the results of which are presented in Fig. 6, right panel and Table 5, right part. The plot shows that the contribution by null connections indeed shrunk significantly, and thus the amount of classification errors was decreased almost by half. At the same time, for some neurons *α*=10 turned out to be too strong of a regularization factor, and thus the estimator, in an overzealous attempt to find a sparse solution, set to zero some of the weaker excitatory and inhibitory synapse weights. This can be seen as a secondary peak of the red distribution at the origin. A secondary peak of the blue distribution is also present, but scarcely visible due to scale.

### 3.4 Synfire chain embedded in a random balanced network

#### 3.4.1 Construction of the network model

In this experiment, we turned to structured networks in order to highlight the generative aspects of the proposed GLM model and demonstrate a potential approach to the interpretation of the recovered connectivity. One specific structure of interest, prominent in the context of cortical networks, is called a “synfire chain” (Abeles 1982). The synfire chain, consisting of consecutively linked and synchronously activated groups of neurons, is a thoroughly studied model of signal propagation in the cortex (Diesmann et al. 1999; Goedeke and Diesmann 2008).

We built a simulation of a random balanced network with an embedded synfire chain, simulated the dynamics of this network and recorded its spiking activity, which we then used as input data for the MLE procedure to infer the parameters of our GLM (no regularization was applied in this experiment, unlike in the last case presented in Section 3.3). However, as would be the case with the experimental recordings, we did not assume that we know the “right” ordering of the neuron identifiers. We therefore subjected the recovered connectivity to a clustering process in order to reveal the trace of the synfire chain in the connection matrix. After identifying the synfire chain in the network, we performed a simulation where we stimulated the discovered first “link” of the chain in the original and reconstructed networks, and observed identical dynamics in both cases.

Similarly to the previous experiments, we first constructed a random balanced network of LIF neurons (*N*=1000) with 80 % : 20 % proportion of excitatory to inhibitory cells. This time, we used “iaf_psc_delta_canon” model in NEST nomenclature; this model is different from the standard “iaf_psc_delta” and “iaf_psc_alpha” LIF neurons in that the points in time when it emits spikes are not tied to the grid defined by the simulation resolution, but rather are recorded precisely as they occur (Morrison et al. 2007; Hanuschkin et al. 2010). Correspondingly, for the external inputs, we employed the continuous time version of the Poisson generator “poisson_generator_ps”. Since this network model works in continuous time and does not require discretization or binning of the spike data, we wanted to examine the implications of feeding the precise spike times to the MLE of the GLM, as opposed to data binned to Δ*t*=0.1 ms simulation resolution as in the previous experiments. The model parameters were fixed to *τ* _{m}=20 ms, *τ* _{r}=2 ms, *𝜃*=20 mV, and *V* _{r}=0 mV. Each neuron was set to receive a fixed number of incoming connections (*M* _{e}=80 excitatory and *M* _{i}=20 inhibitory), where the pre-synaptic neurons were randomly selected (without replacement) from the excitatory and inhibitory populations respectively (implemented as “RandomConvergentConnect” function in NEST). Synaptic weights were set to *J* _{e}=0.9 mV for excitatory, *J* _{i}=−4.5 mV for inhibitory connections with a transmission delay of *d*=1.5 ms. Additional independent Poisson process excitatory inputs were supplied at \(\nu _{\mathrm {e}}=2222\text { s}^{-1}\) and inhibitory inputs at \(\nu _{\mathrm {i}}=0.25\times \nu _{\mathrm {e}}=556\text { s}^{-1}\).

On top of this “background” network, we selected *N* _{l}=10 groups (links) of \(N^{\propto }=50\) neurons each (\(N_{\mathrm {e}}^{\propto }=40\) excitatory and \(N_{\mathrm {i}}^{\propto }=10\) inhibitory cells) and connected all \(N_{\mathrm {e}}^{\propto }\) excitatory neurons of every group to each of the \(N^{\propto }\) neurons in the next group with \(J_{\mathrm {e}}^{\propto }=1.4\text { mV}\) excitatory synapses (transmission delay *d*=1.5 ms). Inhibitory neurons in a link of the chain do not have specific connections to the next link in the chain (Hayon et al. 2004). No neuron in the network was part of more than one group of the synfire chain. This way, we created a “hidden” embedded synfire chain, which receives inputs from the background random network and likewise projects outgoing connections to the background network. When the first group of this structure is stimulated in a coordinated fashion, the chain reliably propagates the excitation from one group to the next until it reaches the last one, and terminates. In the absence of such coordinated stimulation, the synfire chain did not activate, and only “background” activity was observed.

#### 3.4.2 Identification of the synfire chain by connectivity clustering

The complete network was simulated for *T*=2 hours of biological time and exhibited an average firing rate of *ν*=1.4 s^{−1}. The synfire chain was not stimulated during the simulation, so the spike train recordings contained no instances of propagating synfire activity. The neuron identifiers were randomly shuffled and the resulting spike raster was fed into the MLE reconstruction procedure.

We reasoned that one of the most generic differentiators between the neurons that belong to various groups (inhibitory neurons and excitatory neurons that are, or are not part of the synfire chain) is the relative strengths of the synapses (both incoming and outgoing connections can be considered). Therefore, we can apply a clustering algorithm to the recovered connectivity matrix to discern between several classes of neurons. However, most algorithms (such as *k*-means or GMM, employed in the previous sections) require the desired number of clusters to be set explicitly, either through prior knowledge, or by applying statistical or information theory methods to the data to get an estimation.

To circumvent this problem, we carried out an unsupervised learning technique known as hierarchical clustering. It amounts to iteratively repeating the procedure of looking at the discovered clusters (which, in the first step, each contain a single element), determining the ones that are most similar according to a chosen metric, and merging them into an agglomerate cluster; the process continues until a single cluster remains. The results are visualized by constructing a so-called “dendrogram”, which shows the discovered hierarchy of clusters as a tree structure. Therefore, it is not necessary to specify the number of clusters in advance, but rather the most appropriate set of clusters can be selected by analyzing the dendrogram after performing the clustering. This approach fits very well to an exploratory setting, where one might wish to appreciate the entirety of possible groupings in a compact graphical form and then choose the one that best highlights the particular aspect of interest of the data.

We applied hierarchical clustering to the connectivity matrix using Ward’s minimum variance method (Ward 1963) as a criterion for choosing the pair of clusters to merge at each step. Ward’s minimum variance criterion minimizes the total within-cluster variance and enables the grouping of items into sets such that they are maximally similar to each other according to some definition of similarity, which is usually expressed in form of a “dissimilarity matrix”. We used the SciPy hierarchical clustering package (scipy.cluster.hierarchy) to obtain the linkage and visualize the results.

*N*

_{l}=10 links, because the outgoing connections of the last link are not statistically different from those of the background neurons.

The square missing from Fig. 8a is the last link of the chain, which by construction cannot be detected via clustering by the outgoing connections. Therefore, we subjected the neurons that are part of the big yellow cluster (excitatory neurons, which have not been previously identified as taking part in any of the synfire chain links) to additional clustering by incoming connections. This operation reveals the formerly concealed last link of the chain (Fig. 8b). Finally, we applied the same procedure to the inhibitory neurons in the big green cluster. This reveals the inhibitory neurons that are part of the synfire chain. These neurons receive connections from the previous link in the chain but do not send outgoing projections to the next links, and so they are also impossible to detect by clustering only by outgoing connections. This step completes the clustering procedure and we arrive at the final result as shown in Fig. 8c.

The synfire chain is not apparent in the connectivity matrix in randomized ordering, neither for the ground truth matrix (Fig. 9a), nor the MLE-estimated connectivity (Fig. 9d). However, clustering neurons by the similarity of incoming and outgoing connection weights reveals the synfire chain substructure (Fig. 9b, e) of both excitatory and inhibitory neurons. This shows that our clustering procedure successfully recovers the group structure of the synfire chain network. Note that in the original ordering (Fig. 9f), the reconstructed matrix also resembles the ground truth matrix to a great extent (Fig. 9c), as expected based on our previous reconstruction experiments above.

#### 3.4.3 Comparison to correlation-based connectivity estimation

*ρ*

_{ i j }(

*τ*) for all pairs of neurons with a bin size of Δ=10 ms and a maximum time lag of \(\tau _{\max }=\pm 200\text { ms}\). The normalized Pearson cross-correlation coefficient for a stationary ergodic point process for sufficiently large number of sampled bins

*K*is defined as follows (Shao and Chen 1987):

*i*and

*j*(both

*K*-bins long), whereas

*N*

_{ i }and

*N*

_{ j }are the total numbers of spikes of the respective neurons. For each

*ρ*(

*τ*) curve, we found the absolute extrema \(\tau _{\text {peak}}^{ij}=\text {argmax}_{\tau }(|\rho _{ij}(\tau )|)\) and represented the results as a matrix of lagged cross-correlation coefficients \(\tilde {J}_{ij}=\rho _{ij}(\tau _{\text {peak}}^{ij})\), shown in Fig. 9g–i.

We performed clustering on the \(\tilde {J}_{ij}\) matrix as previously described, however, we had to limit ourselves to the first step only, because the cross-correlation matrix is symmetric by construction. The matrix shows similar patterns to the ground truth and MLE connectivity matrices, albeit with substantially lower contrast. Additionally, the direction of the synfire chain cannot be detected, due to the symmetry of the measure mentioned above. Moreover, the individual values of the correlation matrix are difficult to directly relate to the experimental quantities and/or model parameters, because the correlation matrix alone does not constitute a generative model, as we discuss below.

#### 3.4.4 Simulation of original and reconstructed synfire chain networks

Finally, we compared the dynamics of the original and reconstructed network in simulation, including occasional stimulation of the first group of the synfire chain. In both networks, we can identify the order of the groups of the chain by following the links backwards starting from the last link identified in Fig. 8b. Note that the identification of the last link is not determined by the clustering algorithm but simply by membership of the neurons as pre- and post-synaptic partners in the strong connections represented as red boxes in the clustered connectivity matrix. Neurons in the last link occur just as post-synaptic targets; there is no red box in which they occur as pre-synaptic sources. Conversely the neurons of the first link only occur as pre-synaptic sources. Thus the chain can be unrolled from either end by analogous processes.

The simulation of the reconstructed GLM network was carried out in NEST using “pp_psc_delta” neurons and the recovered connectivity matrix. In order to avoid the necessity of fine-tuning the parameters of the stimulation, we additionally included a reset of the membrane potential \(U_{i}(t)\leftarrow 0\) after spike emission (option “with_reset” in the “pp_psc_delta” model, enabled for all neurons), which prevents runaway excitation of the neurons in the network upon delivering a strong stimulus to the synfire chain.

## 4 Discussion

In the present work, we introduce a method for analysis of parallel spike trains based upon maximum likelihood estimation of parameters of a recurrent network of stochastic generalized linear model neurons. The method not only makes it possible to perform large-scale reconstruction of the directed synaptic connectivity of neuronal circuits, but also to recover neuronal parameters, which can be used to obtain a dynamic (i.e. simulatable) model of the network under investigation. Through radical simplification of the single neuron model and interaction kernels as compared to previous studies (Song et al. 2013; Citi et al. 2014; Ramirez and Paninski 2014), the numerics in our method lend themselves to an efficient implementation on both CPUs and GPUs. Moreover, the estimation procedure is highly amenable to parallelization, which makes it possible to scale up the number of units and putative connections dramatically.

The proposed estimation procedure operates in continuous time on precise timestamps of the events (spikes), and does not require discretization, binning or smoothing of the data, which avoids the associated choice of bin or kernel size and induced artifacts (Ba et al. 2014). Additionally, unlike pairwise methods such as the coupled escape rate model (CERM) by Kobayashi and Kitano (2013), the reconstruction takes into account the complete ensemble spike history and thus is able to disambiguate complex indirect neural interactions. Other recently proposed connectivity reconstruction methods, not based on GLMs, exploit specific properties of leaky integrate-and-fire neurons (Van Bussel et al. 2011; Memmesheimer et al. 2014) or of linearly interacting point processes (Pernice and Rotter 2013). While this might be less clear for these methods, our procedure, since it is a MLE of a GLM, can be shown to have the optimality properties of becoming an asymptotically unbiased, consistent and efficient estimator of the ground truth connectivity in the limit of large sample sizes (Pawitan 2001) (provided that the suggested model is appropriate to describe neuronal dynamics). Moreover, it is amenable to efficient optimization via gradient ascent, since it is mathematically guaranteed to converge to the global maximum of the likelihood.

We present benchmarks against simulated random balanced networks of *N*=1000 neurons with known ground truth connectivity, and show that our method achieves good performance for realistic model parameters and plausible amounts of data. Additionally, we performed a successful reconstruction of a structured network, where a synfire chain was embedded in a balanced network of excitatory and inhibitory neurons. The simulation of the reconstructed network with stimulation applied to the first link of the synfire chain, which was identified by carrying out cluster analysis of the recovered synaptic connectivity matrix, highlighted the generative properties of the GLM and showed virtually identical network dynamics to the original network. The application of cluster analysis to the reconstructed connectivity of the synfire chain is an example of how an inferred network model can be subdivided into interacting populations of neurons. Given such a partition of the network in functional subgroups, the activity dynamics can be analyzed using theory of population dynamics of GLM neurons (Deger et al. 2014).

Ideally our approach would also be validated against experimental data. Unfortunately, no datasets are currently available that contain long recordings of many individual spike trains and also the connectivity between the neurons. Indeed, generating such a dataset, although now technically possible, for example, using a high density microelectrode array setup (Ballini et al. 2013), would require extraordinary investment from an experimental laboratory. It is therefore more realistic to hope that the experimental validation of our technique can take place opportunistically on a dataset that is obtained for some other purpose.

In spite of the apparent simplicity of our model, the point process GLM framework that we used is very flexible and can be readily extended with additional features. The exponential kernels that we chose to describe the membrane filtering and nonlinear properties of the neurons can be replaced with more elaborate ones. For instance, previous works have represented neuronal interaction kernels by cosine “bumps” (Pillow et al. 2005), or composition of basis functions, such as Laguerre polynomials or B-splines (Song et al. 2013). However, in order to enable the reconstruction of networks of thousands of units, the key is to use functions that can both guarantee the concavity of the likelihood, as discussed in Paninski (2004), and at the same time make it possible to find analytical closed forms for the resulting expressions to enable efficient evaluation. These considerations, and the notion that the exponential PSP is a coarse first-order approximation to the dynamics of synaptic transmission, were the primary motivations for us to adopt the exponential kernels in this work. However, in Section 3.2 we have demonstrated that this simplification does not affect the reconstruction performance for the data generated by a more complex and realistic LIF model with *α*-shaped PSCs. Besides, we argue that our model would be most useful to investigate network effects, as opposed to the effects explained by intricate features of the dynamics of individual synapses, for which purpose, conversely, smaller-scale, but more detailed models like the one by Song et al. (2013) might be more appropriate. Apart from that, it is possible to add supplementary terms to the membrane potentials of the neurons *U* _{ i }(*t*). One such option is to incorporate known external inputs directly into the model, such as those occurring in experimental paradigms widely used for studying predominantly stimulus-driven circuits like the retina (Pillow et al. 2008). Another option is to add unknown, common external inputs (Kulkarni and Paninski 2007; Vidne et al. 2012) in order to treat non-stationarity in the data.

A further possibility to improve the results of the estimation lies in enforcing Dale’s law: neurons can be either inhibitory or excitatory, and they cannot form connections of both types at the same time (Eccles 1976). Unfortunately, the mathematical re-formulation of this law in the context of our model (the sign of all elements in each column of the synaptic weight matrix *J* _{ i j } should be identical) turns the original problem into a non-convex and non-separable one. Instead of trying to solve this much more difficult optimization problem, an approximate, greedy method can be implemented as outlined in Mishchenko et al. (2011). This involves first solving the original problem, then classifying the neurons as excitatory, inhibitory or unassigned based on the discovered synaptic weights, and, finally, imposing corresponding box constraints on the relevant elements of the *J* _{ i j } rows, which neither compromises the convexity, nor the separability properties. However, in our case, the major source of errors is the overlap between the unconnected and excitatory distributions, which generates non-Dale connections as a consequence. The benchmarks that we conducted show that very few of the errors are of the non-Dale category (see Tables 2, 3, 5), so any gain from imposing a Dale condition would be minimal and does not justify the additional complexity incurred. Therefore, effort should primarily be focused on tightening distributions.

In our GLM, we have used the exponential link function to map the membrane potential *U* _{ i }(*t*) to the instantaneous firing rate *λ* _{ i }(*t*). The exponential function is the canonical link function for the Poisson distribution, and it is commonly used in the single neuron modelling context, e.g. in the spike response model (Gerstner et al. 2014). Further reasons for us to choose an exponential function as the link function were as follows: 1) it has been previously shown (Jolivet et al. 2006; Mensi et al. 2012) that an exponential function is a good model for the nonlinear relationship between the conditional intensity of spike emission and the distance from the voltage threshold; 2) an exponential nonlinearity satisfies the sufficient condition established in Paninski (2004) for the likelihood of the model to be concave; 3) this choice makes it possible for us to obtain the closed form for the likelihood function as an exponential integral Ei (10), which is crucial here for reasons of computational efficiency. If the closed form cannot be obtained, then one either needs to discretize the likelihood integral, possibly using clever corrections to improve the accuracy (Citi et al. 2014), or, if the conditional intensity function can be shown to be piecewise smooth like in our case, attempt to get better precision and performance by applying quadrature methods to the smooth segments (Mena and Paninski 2014). Unfortunately, both approaches are still not fast enough for large GLMs such as ours. Other link functions such as logit and probit functions are also commonly used in the context of GLMs and have the property of being bounded (Song et al. 2013). Indeed, within our framework, it is possible to adopt a different link function instead of the canonical one. However, both logit and probit functions in particular are ruled out by the concavity condition (Paninski 2004), being saturating (“sigmoidal”) nonlinearities. In practice, however, we did not experience any substantial issues due to the exponential function being positively unbounded. The box constraints that we imposed on the base rate and synaptic weights served only to repel the optimizer from the borders of the feasible region, where it might occasionally find itself due to an unfortunate combination of numerical artifacts. In none of the results presented in the paper did the recovered parameters turn out to be equal to the values of the box constraints.

Throughout this study, we have assumed that we have simultaneous access to all the spike trains of a neuronal population. For this case, and for a small number of neurons, it was shown that connectivity estimation via GLM can recover anatomical connectivity (Gerhard et al. 2013), as opposed to other methods, which mostly uncover “functional” or “effective” connections (Stevenson et al. 2008) that do not necessarily correspond to real synapses. Here, we scaled the GLM approach up to large networks. However, in many experiments, such as in cortical multi-electrode array recordings (Truccolo et al. 2010), a complete recording of all neurons in a network is not feasible, but rather only parts of a neuronal network can be observed. With respect to the inference of connectivity from activity this is known as the problem of undersampling (Kim et al. 2011; Gerhard et al. 2011; Shimazaki et al. 2012; Lütcke et al. 2013): an unobserved neuron might excite several observed ones reliably and frequently. Even if these observed neurons are not synaptically connected to each other, connectivity inference methods that do not account for hidden units would infer connections among them to explain the correlations in their activity. Thus, we generally expect the reconstruction accuracy of our method to decrease in case of undersampling of the network, as the input from unobserved units will be “explained” by non-existing connections (see also Appendix C.2). Other experimental preparations, such as neuronal cultures on substrate-integrated multi-electrode arrays, are amenable to more complete recordings (Ballini et al. 2013), possibly enabling direct interpretation of the recovered connectivity.

We emphasize that our method is practical for networks of up to thousands of neurons, and yet we recognize that the machines featuring a large number of cores (>10^{5}), such as the ones we used during the development phase of this project, are generally only to be found at major research institutions. These supercomputing facilities are becoming increasingly available to neuroscience researchers. For example, researchers based in Germany may take advantage of the twice-yearly calls for applications for computing time on the supercomputers at Jülich Supercomputing Centre^{7}, at no cost to the researcher if accepted. European researchers outside Germany can apply analogously for resources through PRACE,^{8} and labs based in the US can apply for time at the NSF facilities.^{9} Additionally, initiatives such as the Human Brain Project^{10} and the Neuroscience Gateway^{11} aim to make such resources more accessible to the neuroscience community. Even so, core-hour allocations often require a thorough justification and quantitative evidence of the scaling properties of the algorithm, both of which entail significant investment from the researcher in preparing the application.

Therefore, we also investigated the option of offloading the computations to the kinds of GPGPU accelerators that are currently available off-the-shelf. We implemented a naive version of a GPU port, in which the computation kernels originally written in C++ and parallelized using OpenMP to use multiple threads were rewritten using CUDA technology by Nvidia Corporation to use a GPU instead. In order to assess the performance of this port, we measured the time it takes to complete the reconstruction of the incoming synapses of one neuron of a network of a thousand of neurons, such as those presented in Section 3. Both applications were tested on an IBM System x iDataPlex dx360 M4 machine featuring two Intel Xeon X5650 processors (6 cores, 12 threads) and one NVIDIA Tesla M2070 (Fermi microarchitecture). The CPU version took 38 minutes to converge in 433 iterations, while the GPU port required 49 minutes and 427 iterations; the obtained log-likelihood values were identical up to an absolute difference of ≃4×10^{−10} and a relative difference of ≃3×10^{−15}. This way, the speedup achieved by offloading the calculations to a single GPU as compared to a single CPU thread amounted to approximately a factor of 18. However, profiling revealed, that around 70 % of the runtime of the GPU port was not actually spent doing useful calculations, but rather transferring *ν* _{ i j } vectors from the CPU to the GPU memory. Therefore, simply switching to a better GPU, such as the ones based on the Kepler microarchitecture, providing double of the data transfer bandwidth as compared to Fermi, will increase the speedup for a naive GPU port up to a factor of 28. Furthermore, we are currently investigating algorithmic improvements that completely remove the need for data transfers by storing *ν* _{ i j } vectors directly in the GPU memory using specialized compression. Extrapolating on the performance from the proof-of-concept kernels we implemented, a future GPU realization may perform at least as fast as ∼55 generic CPU threads, and require only a fraction of RAM as compared to the CPU-only realization by storing all of the working data in the onboard GPU memory. The complete GPU port of the method, along with its core algorithms and performance benchmarks, will be described in detail in a separate publication. The development of a substantially more efficient implementation will enable us to thoroughly investigate the limits of our approach in a way that is out of scope in the current study due to computational expense. One obvious area for investigation would be the degradation of performance in the case of undersampling as discussed above. Other areas worthy of further examination are the effects on misclassification error rates of correlated external inputs and non-stationarities in the recorded activity.

It is also important to mention that anatomically, cortical neurons receive on the order of 10^{3}- 10^{4} incoming synapses (Braitenberg and Schüz 1991). In our demonstrations, we assumed that the network might be fully connected, or, in other words, each of the *N*=1000 neurons can possibly receive up to 10^{3} incoming synapses from all other neurons, yielding \(\mathcal {O}(N^{2}=10^{6})\) parameters to constrain in total. However, given a substantially larger amount of recording channels, such as *N*>10^{4}, if such data becomes available, this assumption is no longer reasonable. Instead, the data can be pre-processed to purge unlikely incoming connections, from *N* down to the most likely 10^{3}- 10^{4} putative synapses per neuron, thereby avoiding the quadratic explosion in the number of model parameters. We suggest that such pre-processing can be performed using computationally efficient pairwise linear methods, such as cross-correlation or cross-coherence, or various information theory metrics (Staniek and Lehnertz 2008). This way, while recovering the connectivity of even larger networks would still require a linear increase in computational resources (or wallclock time), the challenge to further scale the model up to a larger number of putative incoming synapses can be alleviated.

Finally, we would like to stress that even though network models that can be directly simulated as extracted from the data are interesting in themselves for further studies, the proposed method also has potential to provide insights into the network-wide plasticity of synaptic connectivity. Even though in our method we assume that the connectivity is fixed over the time of a recording, estimated synaptic weights can be tracked accross several recordings performed in a time-lapse fashion. Such data could be relevant for models of synaptic plasticity over long time-scales (structural plasticity) (Escobar et al. 2008; Deger et al. 2012), which currently have to rely on statistics of synapse numbers without temporal information, or time-lapse imaging of small numbers of individual synapses.

## Footnotes

## Notes

### Acknowledgments

We thank Stefano Cardanobile, Stefan Rotter, Wolfram Schenck and Christian Pozzorini for helpful discussions, and Ulrich Egert for his support of the initial phase of the project. We are grateful to Andrew V. Adinetz for his advice on analyzing the performance of our optimizer and on the GPU kernel implementation, and to Tom Tetzlaff for providing an example synfire chain simulation script. We acknowledge partial support by the German Federal Ministry of Education and Research (BMBF grants 01GQ0420 to BCCN Freiburg and 01GQ0830 to BFNT Freiburg), the Swiss National Science Foundation (grant agreement no. 200020_147200), and the Helmholtz Alliance through the Initiative and Networking Fund of the Helmholtz Association and the Helmholtz Portfolio theme “Supercomputing and Modeling for the Human Brain”.

### Conflict of interests

The authors declare that they have no conflict of interest.

### Open Access

This article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.

## Supplementary material

## References

- Abeles, M. (1982).
*Local cortical circuits: an electrophysiological study, Studies of brain function*Vol. 6. Berlin: Springer. doi: 10.1007/978-3-642-81708-3. - Aertsen, A.M.H.J., Gerstein, G.L., Habib, M.K., & Palm, G. (1989). Dynamics of neuronal firing correlation: modulation of “effective connectivity”.
*Journal of Neurophysiology*,*61*(5), 900–917.PubMedGoogle Scholar - Ahrens, M.B., Orger, M.B., Robson, D.N., Li, J.M., & Keller, P.J. (2013). Whole-brain functional imaging at cellular resolution using light-sheet microscopy.
*Nature Methods*,*10*(5), 413–420. doi: 10.1038/nmeth.2434.PubMedCrossRefGoogle Scholar - Ba, D., Temereanca, S., & Brown, E.N. (2014). Algorithms for the analysis of ensemble neural spiking activity using simultaneous-event multivariate point-process models.
*Frontiers in Computational Neuroscience*,*8*, 6. doi: 10.3389/fncom.2014.00006.PubMedCentralPubMedCrossRefGoogle Scholar - Ballini, M., Muller, J., Livi, P., Chen, Y., Frey, U., Shadmani, A., Jones, I., Gong, W., Fiscella, M., Radivojevic, M., Bakkum, D., Stettler, A., Heer, F., & Hierlemann, A. (2013). A 1024-channel CMOS microelectrode-array system with 26’400 electrodes for recording and stimulation of electro-active cells in-vitro. In
*Symposium on VLSI Circuits (VLSIC)*.Google Scholar - Bock, D.D., Lee, W.C.A., Kerlin, A.M., Andermann, M.L., Hood, G., Wetzel, A.W., Yurgenson, S., Soucy, E.R., Kim, H.S., & Reid, R.C. (2011). Network anatomy and in vivo physiology of visual cortical neurons.
*Nature*,*471*(7337), 177–182. doi: 10.1038/nature09802.PubMedCentralPubMedCrossRefGoogle Scholar - Boucsein, C., Nawrot, M.P., Schnepel, P., & Aertsen, A. (2011). Beyond the cortical column: abundance and physiology of horizontal connections imply a strong role for inputs from the surround.
*Frontiers in Neuroscience*,*5*, 32. doi: 10.3389/fnins.2011.00032.PubMedCentralPubMedCrossRefGoogle Scholar - Braitenberg, V., & Schüz, A. (1991).
*Anatomy of the cortex: Statistics and geometry, Studies of brain function*Vol. 18. Berlin: Springer.Google Scholar - Briggman, K.L., Helmstaedter, M., & Denk, W. (2011). Wiring specificity in the direction-selectivity circuit of the retina.
*Nature*,*471*(7337), 183–188. doi: 10.1038/nature09818.PubMedCrossRefGoogle Scholar - Brillinger, D.R. (1988). Maximum likelihood analysis of spike trains of interacting nerve cells.
*Biological Cybernetics*,*59*(3), 189–200.PubMedCrossRefGoogle Scholar - Brunel, N. (2000). Dynamics of sparsely connected networks of excitatory and inhibitory spiking neurons.
*Journal of Computational Neuroscience*,*8*(3), 183–208. doi: 10.1023/A:1008925309027.PubMedCrossRefGoogle Scholar - Byrd, R.H., Lu, P., Nocedal, J., & Zhu, C. (1995). A limited memory algorithm for bound constrained optimization.
*SIAM Journal on Scientific Computing*,*16*(5), 1190–1208 . doi: 10.1137/0916069.CrossRefGoogle Scholar - Chen, Z., Putrino, D.F., Ghosh, S., Barbieri, R., & Brown, E.N. (2011). Statistical inference for assessing functional connectivity of neuronal ensembles with sparse spiking data.
*IEEE Transactions on Neural Systems and Rehabilitation Engineering*,*19*(2), 121–135. doi: 10.1109/TNSRE.2010.2086079.PubMedCentralPubMedCrossRefGoogle Scholar - Chu, C., Kim, S.K., Lin, Y., Yu, Y., Bradski, G.R., Ng, A.Y., & Olukotun, K. (2006). Map-reduce for machine learning on multicore. In Schölkopf, B., Platt, J., & Hoffman, T. (Eds.)
*Advances in Neural Information Processing Systems 19, Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems*(pp. 281–288). British Columbia, Canada: MIT Press, Vancouver. http://papers.nips.cc/paper/3150-map-reduce-for-machine-learning-on-multicore. - Chung, K., Wallace, J., Kim, S.Y., Kalyanasundaram, S., Andalman, A.S., Davidson, T.J., Mirzabekov, J.J., Zalocusky, K.A., Mattis, J., Denisin, A.K., Pak, S., Bernstein, H., Ramakrishnan, C., Grosenick, L., Gradinaru, V., & Deisseroth, K. (2013). Structural and molecular interrogation of intact biological systems.
*Nature*,*497*(7449), 332–337. doi: 10.1038/nature12107.PubMedCentralPubMedCrossRefGoogle Scholar - Citi, L., Ba, D., Brown, E.N., & Barbieri, R. (2014). Likelihood methods for point processes with refractoriness.
*Neural Computation*,*26*(2), 237–263. doi: 10.1162/NECO_a_00548.PubMedCrossRefGoogle Scholar - Cody, W., & Thacher, H.C. (1969). Chebyshev approximations for the exponential integral Ei(
*x*).*Mathematics of Computation*,*23*(106), 289–303. http://www.ams.org/mcom/1969-23-106/S0025-5718-1969-0242349-2/.Google Scholar - Deger, M., Helias, M., Rotter, S., & Diesmann, M. (2012). Spike-timing dependence of structural plasticity explains cooperative synapse formation in the neocortex.
*PLoS Computational Biology*,*8*(9), e1002,689. doi: 10.1371/journal.pcbi.1002689.CrossRefGoogle Scholar - Deger, M., Schwalger, T., Naud, R., & Gerstner, W. (2014). Fluctuations and information filtering in coupled populations of spiking neurons with adaptation.
*Physical Review E*,*90*, 062,704. doi: 10.1103/PhysRevE.90.062704.CrossRefGoogle Scholar - Diesmann, M., Gewaltig, M.O., & Aertsen, A. (1999). Stable propagation of synchronous spiking in cortical neural networks.
*Nature*,*402*(6761), 529–533. doi: 10.1038/990101.PubMedCrossRefGoogle Scholar - Eccles, J. (1976). From electrical to chemical transmission in the central nervous system.
*Notes and Records of the Royal Society of London*,*30*(2), 219–230.PubMedCrossRefGoogle Scholar - Escobar, G., Fares, T., & Stepanyants, A. (2008). Structural plasticity of circuits in cortical neuropil.
*The Journal of Neuroscience*,*28*(34), 8477–8488. doi: 10.1523/JNEUROSCI.2046-08.2008.PubMedCentralPubMedCrossRefGoogle Scholar - Gerhard, F., Pipa, G., Lima, B., Neuenschwander, S., & Gerstner, W. (2011).
*Frontiers in Computational Neuroscience*,*5*, 4. doi: 10.3389/fncom.2011.00004.PubMedCentralPubMedCrossRefGoogle Scholar - Gerhard, F., Kispersky, T., Gutierrez, G.J., Marder, E., Kramer, M., & Eden, U. (2013). Successful reconstruction of a physiological circuit with known connectivity from spiking activity alone.
*PLoS Computational Biology*,*9*(7), e1003,138. doi: 10.1371/journal.pcbi.1003138.CrossRefGoogle Scholar - Gerstein, G.L., & Perkel, D.H. (1969). Simultaneously recorded trains of action potentials: analysis and functional interpretation.
*Science*,*164*(881), 828–830. doi: 10.1126/science.164.3881.828.PubMedCrossRefGoogle Scholar - Gerstner, W., Kistler, W.M., Naud, R., & Paninski, L. (2014).
*Neuronal Dynamics: From Single Neurons to Networks and Models of Cognition*. New York: Cambridge University Press.Google Scholar - Gerwinn, S., Macke, J.H., & Bethge, M. (2010). Bayesian inference for generalized linear models for spiking neurons.
*Frontiers in Computational Neuroscience*,*4*, 12. doi: 10.3389/fncom.2010.00012.PubMedCentralPubMedCrossRefGoogle Scholar - Gewaltig, M.O., & Diesmann, M. (2007). NEST (NEural Simulation Tool).
*Scholarpedia*,*2*(4), 1430.CrossRefGoogle Scholar - Ghane-Motlagh, B., & Sawan, M. (2013). Design and implementation challenges of microelectrode arrays: a review.
*Materials Sciences and Applications*,*4*(08), 483. doi: 10.4236/msa.2013.48059.CrossRefGoogle Scholar - Goedeke, S., & Diesmann, M. (2008). The mechanism of synchronization in feed-forward neuronal networks.
*New Journal of Physics*,*10*(1), 015,007. doi: 10.1088/1367-2630/10/1/015007.CrossRefGoogle Scholar - Grewe, B.F., & Helmchen, F. (2009). Optical probing of neuronal ensemble activity.
*Curr. Opin. Neurobiol.*,*19*(5), 520–529. doi: 10.1016/j.conb.2009.09.003.PubMedCrossRefGoogle Scholar - Hanuschkin, A., Kunkel, S., Helias, M., Morrison, A., & Diesmann, M. (2010). A general and efficient method for incorporating precise spike times in globally time-driven simulations.
*Front. Neuroinform.*,*4*, 113. doi: 10.3389/fninf.2010.00113.PubMedCentralPubMedGoogle Scholar - Hatsopoulos, N.G., & Donoghue, J.P. (2009). The science of neural interface systems.
*Annual Review of Neuroscience*,*32*, 249–266. doi: 10.1146/annurev.neuro.051508.135241.PubMedCentralPubMedCrossRefGoogle Scholar - Hayon, G., Abeles, M., & Lehmann, D. (2004). Modeling compositionality by dynamic binding of synfire chains.
*Journal of Computational Neuroscience*,*17*(2), 179–201. doi: 10.1023/B:JCNS.0000037682.18051.5f.PubMedCrossRefGoogle Scholar - Horwitz, B. (2003). The elusive concept of brain connectivity.
*Neuroimage*,*19*(2 Pt 1), 466–470. doi: 10.1016/S1053-8119(03)00112-5.PubMedCrossRefGoogle Scholar - Jolivet, R., Rauch, A., Lüscher, H.R., & Gerstner, W. (2006). Predicting spike timing of neocortical pyramidal neurons by simple threshold models.
*Journal of Computational Neuroscience*,*21*(1), 35–49. doi: 10.1007/s10827-006-7074-5.PubMedCrossRefGoogle Scholar - Kim, S., Putrino, D., Ghosh, S., & Brown, E.N. (2011). A Granger causality measure for point process models of ensemble neural spiking activity.
*PLoS Computational Biology*,*7*(3), e1001,110. doi: 10.1371/journal.pcbi.1001110.CrossRefGoogle Scholar - Kobayashi, R., & Kitano, K. (2013). Impact of network topology on inference of synaptic connectivity from multi-neuronal spike data simulated by a large-scale cortical network model.
*Journal of Computational Neuroscience*,*35*(1), 109–124. doi: 10.1007/s10827-013-0443-y.PubMedCrossRefGoogle Scholar - Kohavi, R. (1995). A study of cross-validation and bootstrap for accuracy estimation and model selection. In
*International Joint Conference on Artificial Intelligence*, (Vol. 14 pp. 1137–1145).Google Scholar - Kulkarni, J.E., & Paninski, L. (2007). Common-input models for multiple neural spike-train data.
*Network*,*18*(4), 375–407. doi: 10.1080/09548980701625173.PubMedCrossRefGoogle Scholar - Lee, S.I., Lee, H., Abbeel, P., & Ng, A.Y. (2006). Efficient
*ℓ*_{1}regularized logistic regression. In*Proceedings of the National Conference on Artificial Intelligence*, (Vol. 21 p. 401). Menlo Park, CA; Cambridge, MA; London: AAAI Press; MIT Press; 1999.Google Scholar - Liu, D.C., & Nocedal, J. (1989). On the limited memory BFGS method for large scale optimization.
*Mathematical Programming*,*45*(1-3), 503–528. doi: 10.1007/bf01589116.CrossRefGoogle Scholar - Lütcke, H., Gerhard, F., Zenke, F., Gerstner, W., & Helmchen, F. (2013). Inference of neuronal network spike dynamics and topology from calcium imaging data.
*Front. Neural. Circuits*,*7*, 201. doi: 10.3389/fncir.2013.00201.PubMedCentralPubMedCrossRefGoogle Scholar - McCullagh, P., & Nelder, J.A. (1989).
*Generalized linear models, 2nd edn. No. 37 in Monographs on statistics and applied probability*. London: Chapman and Hall.Google Scholar - Meinshausen, N., & Bühlmann, P. (2006). High-dimensional graphs and variable selection with the Lasso.
*The Annals of Statistics*,*34*(3), 1436–1462. doi: 10.1214/009053606000000281.CrossRefGoogle Scholar - Memmesheimer, R.M., Rubin, R., Olveczky, B.P., & Sompolinsky, H. (2014). Learning precisely timed spikes.
*Neuron*,*82*(4), 925–938. doi: 10.1016/j.neuron.2014.03.026.PubMedCrossRefGoogle Scholar - Mena, G., & Paninski, L. (2014). On quadrature methods for refractory point process likelihoods.
*Neural Computation*,*26*(12), 2790–2797. doi: 10.1162/NECO_a_00676.PubMedCrossRefGoogle Scholar - Mensi, S., Naud, R., Pozzorini, C., Avermann, M., Petersen, C.C.H., & Gerstner, W. (2012). Parameter extraction and classification of three cortical neuron types reveals two distinct adaptation mechanisms.
*Journal of Neurophysiology*,*107*(6), 1756–1775. doi: 10.1152/jn.00408.2011.PubMedCrossRefGoogle Scholar - Mishchenko, Y., Vogelstein, J., & Paninski, L. (2011). A Bayesian approach for inferring neuronal connectivity from calcium fluorescent imaging data.
*Annals of Applied Statistics*,*5*(2B), 1229–1261. doi: 10.1214/09-AOAS303.CrossRefGoogle Scholar - Morrison, A., Straube, S., Plesser, H.E., & Diesmann, M. (2007). Exact subthreshold integration with continuous spike times in discrete-time neural network simulations.
*Neural Computation*,*19*(1), 47–79. doi: 10.1162/neco.2007.19.1.47.PubMedCrossRefGoogle Scholar - Nam, Y., & Wheeler, B.C. (2011). In vitro microelectrode array technology and neural recordings.
*Critical Reviews in Biomedical Engineering*,*39*(1), 45–61. doi: 10.1615/CritRevBiomedEng.v39.i1.40.PubMedCrossRefGoogle Scholar - Oh, S.W., Harris, J.A., Ng, L., Winslow, B., Cain, N., Mihalas, S., Wang, Q., Lau, C., Kuan, L., Henry, A.M., Mortrud, M.T., Ouellette, B., Nguyen, T.N., Sorensen, S.A., Slaughterbeck, C.R., Wakeman, W., Li, Y., Feng, D., Ho, A., Nicholas, E., Hirokawa, K.E., Bohn, P., Joines, K.M., Peng, H., Hawrylycz, M.J., Phillips, J.W., Hohmann, J.G., Wohnoutka, P., Gerfen, C.R., Koch, C., Bernard, A., Dang, C., Jones, A.R., & Zeng, H. (2014). A mesoscale connectome of the mouse brain.
*Nature*,*508*(7495), 207–214. doi: 10.1038/nature13186.PubMedCrossRefGoogle Scholar - Okatan, M., Wilson, M.A., & Brown, E.N. (2005). Analyzing functional connectivity using a network likelihood model of ensemble neural spiking activity.
*Neural Computation*,*17*(9), 1927–1961. doi: 10.1162/0899766054322973.PubMedCrossRefGoogle Scholar - Paninski, L. (2004). Maximum likelihood estimation of cascade point-process neural encoding models.
*Network*,*15*(4), 243–262.PubMedCrossRefGoogle Scholar - Pawitan, Y. (2001).
*In all likelihood: statistical modelling and inference using likelihood*. London: Oxford University Press.Google Scholar - Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cournapeau, D., Brucher, M., Perrot, M., & Duchesnay, E. (2011). Scikit-learn: Machine learning in Python.
*Journal of Machine Learning Research*,*12*, 2825–2830.Google Scholar - Perin, R., Berger, T.K., & Markram, H. (2011). A synaptic organizing principle for cortical neuronal groups.
*Proceedings of the National Academy of Sciences of the United States of America*,*108*(13), 5419–5424. doi: 10.1073/pnas.1016051108.PubMedCentralPubMedCrossRefGoogle Scholar - Pernice, V., & Rotter, S. (2013). Reconstruction of sparse connectivity in neural networks from spike train covariances.
*Journal of Statistical Mechanics: Theory and Experiment*,*2013*(03), P03,008. doi: 10.1088/1742-5468/2013/03/p03008.CrossRefGoogle Scholar - Pillow, J.W., Paninski, L., Uzzell, V.J., Simoncelli, E.P., & Chichilnisky, E.J. (2005). Prediction and decoding of retinal ganglion cell responses with a probabilistic spiking model.
*Journal of Neuroscience*,*2547*, 11,003–11,013. doi: 10.1523/JNEUROSCI.3305-05.2005.CrossRefGoogle Scholar - Pillow, J.W., Shlens, J., Paninski, L., Sher, A., Litke, A.M., Chichilnisky, E.J., & Simoncelli, E.P. (2008). Spatio-temporal correlations and visual signalling in a complete neuronal population.
*Nature*,*454*(7207), 995–999. doi: 10.1038/nature07140.PubMedCentralPubMedCrossRefGoogle Scholar - Ramirez, A.D., & Paninski, L. (2014). Fast inference in generalized linear models via expected log-likelihoods.
*Journal of Computational Neuroscience*,*36*(2), 215–234. doi: 10.1007/s10827-013-0466-4.PubMedCentralPubMedCrossRefGoogle Scholar - Ravikumar, P., Wainwright, M.J., & Lafferty, J.D. (2010). High-dimensional Ising model selection using
*ℓ*_{1}-regularized logistic regression.*Annal of Statistics*,*38*(3), 1287–1319. doi: 10.1214/09-AOS691.CrossRefGoogle Scholar - Schmidt, M., Fung, G., & Rosales, R. (2009).
*Optimization methods for**ℓ*_{1}-*regularization*. Tech. Rep. TR-2009-19, University of British Columbia. http://www.cs.ubc.ca/cgi-bin/tr/2009/TR-2009-19. - Shao, X.S., & Chen, P.X. (1987). Normalized auto- and cross-covariance functions for neuronal spike train analysis.
*The International Journal of Neuroscience*,*34*(1-2), 85–95.PubMedCrossRefGoogle Scholar - Shimazaki, H., Amari, S.I., Brown, E.N., & Grün, S. (2012). State-space analysis of time-varying higher-order spike correlation for multiple neural spike train data.
*PLoS Computational Biology*,*83*, e1002,385. doi: 10.1371/journal.pcbi.1002385.CrossRefGoogle Scholar - Simoncelli, E., Paninski, L., Pillow, J., & Schwartz, O. (2004).
*The New Cognitive Neurosciences*. Cambridge, MA, London, England: MIT Press. chap Characterization of neural responses with stochastic stimuli. Bradford Books.Google Scholar - Snyder, D.L., & Miller, M.I. (1991).
*Random Point Processes in Time and Space (Springer Texts in Electrical Engineering)*. New York: Springer. doi: 10.1007/978-1-4612-3166-0. - Song, D., Wang, H., Tu, C.Y., Marmarelis, V.Z., Hampson, R.E., Deadwyler, S.A., & Berger, T.W. (2013). Identification of sparse neural functional connectivity using penalized likelihood estimation and basis functions.
*Journal of Computational Neuroscience*,*35*(3), 335–357. doi: 10.1007/s10827-013-0455-7.PubMedCentralPubMedCrossRefGoogle Scholar - Spira, M.E., & Hai, A. (2013). Multi-electrode array technologies for neuroscience and cardiology.
*Nature Nanotechnology*,*8*(2), 83–94. doi: 10.1038/nnano.2012.265.PubMedCrossRefGoogle Scholar - Staniek, M., & Lehnertz, K. (2008). Symbolic transfer entropy.
*Physical Review Letters*,*100*(15), 158,101. doi: 10.1103/PhysRevLett.100.158101.CrossRefGoogle Scholar - Stevenson, I.H., Rebesco, J.M., Miller, L.E., & Körding, K.P. (2008). Inferring functional connections between neurons.
*Current Opinion in Neurobiology*,*18*(6), 582–588. doi: 10.1016/j.conb.2008.11.005. - Stevenson, I.H., Rebesco, J.M., Hatsopoulos, N.G., Haga, Z., Miller, L.E., & Körding, K.P. (2009). Bayesian inference of functional connectivity and network structure from spikes.
*IEEE Transactions on Neural Systems and Rehabilitation*,*17*(3), 203–213. doi: 10.1109/TNSRE.2008.2010471. - Tibshirani, R. (1996). Regression shrinkage and selection via the lasso.
*J Roy Stat Soc B Met*, 267–288.Google Scholar - Truccolo, W., Eden, U.T., Fellows, M.R., Donoghue, J.P., & Brown, E.N. (2005). A point process framework for relating neural spiking activity to spiking history, neural ensemble, and extrinsic covariate effects.
*Journal of Neurophysiology*,*93*(2), 1074–1089. doi: 10.1152/jn.00697.2004. - Truccolo, W., Hochberg, L.R., & Donoghue, J.P. (2010). Collective dynamics in human and monkey sensorimotor cortex: predicting single neuron spikes.
*Nature Neuroscience*,*13*(1), 105–111. doi: 10.1038/nn.2455. - Van Bussel, F., Kriener, B., & Timme, M. (2011). Inferring synaptic connectivity from spatio-temporal spike patterns.
*Frontiers in Computational Neuroscience*,*5*, 3. doi: 10.3389/fncom.2011.00003.PubMedCentralPubMedGoogle Scholar - Vidne, M., Ahmadian, Y., Shlens, J., Pillow, J.W., Kulkarni, J., Litke, A.M., Chichilnisky, E.J., Simoncelli, E., & Paninski, L. (2012). Modeling the impact of common noise inputs on the network activity of retinal ganglion cells.
*Journal of Computational Neuroscience*,*33*(1), 97–121 . doi: 10.1007/s10827-011-0376-2.PubMedCentralPubMedCrossRefGoogle Scholar - Ward, J.H. (1963). Hierarchical grouping to optimize an objective function.
*JJournal of the American Statistical Association*,*58*(301), 236–244.CrossRefGoogle Scholar