# Principal component analysis of ensemble recordings reveals cell assemblies at high temporal resolution

- 2k Downloads
- 36 Citations

## Abstract

Simultaneous recordings of many single neurons reveals unique insights into network processing spanning the timescale from single spikes to global oscillations. Neurons dynamically self-organize in subgroups of coactivated elements referred to as cell assemblies. Furthermore, these cell assemblies are reactivated, or replayed, preferentially during subsequent rest or sleep episodes, a proposed mechanism for memory trace consolidation. Here we employ Principal Component Analysis to isolate such patterns of neural activity. In addition, a measure is developed to quantify the similarity of instantaneous activity with a template pattern, and we derive theoretical distributions for the null hypothesis of no correlation between spike trains, allowing one to evaluate the statistical significance of instantaneous coactivations. Hence, when applied in an epoch different from the one where the patterns were identified, (e.g. subsequent sleep) this measure allows to identify times and intensities of reactivation. The distribution of this measure provides information on the dynamics of reactivation events: in sleep these occur as transients rather than as a continuous process.

## Keywords

PCA Reactivation Sleep Memory consolidation## 1 Introduction

*Ensemble* recordings, or the simultaneous recordings of groups of tens to hundreds cells from one or multiple brain areas in behaving animals, offer a valuable window into the network mechanisms and information processing in the brain which ultimately leads to behavior. In the last two decades, the dramatic increase in yield of such techniques with the use of tetrodes, silicon probes and other devices (McNaughton et al. 1983; Buzsáki 2004) poses extremely challenging problems to the data analyst trying to represent and interpret such high-dimensional data and uncover the organization of network activity.

Starting with Donald Hebb’s seminal work (Hebb 1949), theorists have posited *cell assemblies*, or subgroups of coactivated cells, as the main unit of information representation. In this theory, information is represented by *patterns* of cell activity, which create a coherent, powerful input to downstream areas. Cells assemblies would result from modifications of local synapses, e.g. according to Hebb’s rule (Hebb 1949). Their expression and dynamics are likely driven to a large extent by specific interactions between principal cells and interneurons (Geisler et al. 2007; Wilson and Laurent 2005, Benchenane et al. 2008). From an experimental point of view, cell assemblies can be characterized in terms of the coordinated firing of several neurons in a given temporal window, either simultaneously (Harris et al. 2003), or in ordered sequences of action potentials from different cells, as has been shown in both hippocampus (Lee and Wilson 2002) and neocortex (Ikegaya et al. 2004). Ensemble recording provides the opportunity to measure these co-activations in the brain of behaving animals.

To date, only few methods for rigorous statistically based quantification of cell assemblies have been proposed (e.g. Pipa et al. 2008). This problem is all the more delicate when temporal ordering of cells’ discharges is taken into consideration (Mokeichev et al. 2007), requiring immense data sets in order to attain the necessary statistical power (Ji and Wilson 2006). On the other hand, cell assemblies’ zeros-lag co-activations already provide a rich picture of network function (Nicolelis et al. 1995; Riehle et al. 1997), and may represent an easier target for statistical pattern recognition methods. Moreover, the effective connectivity between cells is a dynamical, rapidly changing parameter. Because of this, it is important to follow cell assemblies at a rapid time scale. This would improve our understanding of the temporal evolution of the interaction between cells, their link to brain rhythms, the activity in other brain areas or ongoing behavior.

Principal Component Analysis (PCA) has previously been used to find groups of neurons that tend to fire together in a given time window (Nicolelis et al. 1995; Chapin and Nicolelis 1999). PCA (see e.g. Bishop 1995 can be applied to the correlation matrix of binned multi-unit spike trains, and provides a reduced dimensionality representation of ensemble activity in terms of PC *scores*, i.e., the projections over the eigenvectors of the correlation matrix associated with the largest eigenvalues. This representation accounts for the largest fraction of the variance of the original data for a fixed dimensionality. Also, PCA is intimately related to Hebb’s plasticity rule: it has been shown to be an emergent property of hebbian learning in artificial neural networks (Oja 1982; Bourlard and Kamp 1988).

A remarkable success of ensemble recordings was the demonstration that, during sleep, neural patterns of activity appearing in the immediately previous awake experience are *replayed* (Wilson and McNaughton 1994; Nádasdy et al. 1999; Lee and Wilson 2002; Ji and Wilson 2006). This is proposed to be important for memory consolidation, i.e. turning transient, labile synaptic modifications induced during experience into stable long-term memory traces. Replay appears to take place chiefly during Slow Wave Sleep (SWS). In the hippocampus, a brain structure strongly implicated in facilitating long term memory (Scoville and Milner 1957; Marr 1971; Squire and Zola-Morgan 1991; Nadel and Moscovitch 1997), cell assemblies observed during wakefulness are replayed in subsequent SWS episodes (Wilson and McNaughton 1994) in the form of cell firing sequences (Nádasdy et al. 1999; Lee and Wilson 2002). This occurs during coordinated bursts of activity known as sharp waves (Kudrimoti et al. 1999).

To detect replay, we first need to characterize the activity during active experience, and to generate representative templates from it. Then, templates are compared with the activity during sleep to assess their repetitions. Previous methods have only provided a measure of the overall amount of replay occurring during a whole sleep episode (Wilson and McNaughton 1994; Kudrimoti et al. 1999), by using the session-wide correlation matrix as a template. Alternatively, templates have been generated from the neural activity during a fixed, repetitive behavioral sequence. This is possible, for example, for hippocampal place cells, which fire as the animal follows a trajectory through the firing fields of the respective neurons (Lee and Wilson 2002), if the animal runs through the same trajectory multiple times (Louie and Wilson 2001; Ji and Wilson 2006; Euston et al. 2007), or when a new and transient experience occurs (Ribeiro et al. 2004). One can then search for the repetition of this template during the sleep phase. However, such a template construction technique is not applicable when the behavior is not repeated, or if the behavioral correlates of the recorded cells are not known.

Recently, we used PCA to identify patterns in prefrontal cortical neurons ensembles (Peyrache et al. 2009), without making reference to behavioral sequences, and we devised a novel, simple measure using the PCA-extracted patterns to assess the instantaneous similarity of the activity during sleep. During sleep, this similarity increased in strong transients demonstrating that neuron ensembles appearing in the AWAKE phase reactivate during SWS. Further, the fine temporal resolution of this approach uncovered for the first time a link between assembly replay in the prefrontal cortex and hippocampal sharp waves, as well as the relationship between this replay and slow cortical oscillations. It was also possible to determine the precise behavioral events corresponding to the origin of replayed patterns. Moreover, we were able to determine that the formation of these cell assemblies involves specific interactions between interneurons and pyramidal cells (Benchenane et al. 2008).

The present paper presents this methodology in detail with mathematical and statistical support, and provides new results on how the statistical significance of the replay can be assessed. We show how random matrix theory can be used to provide analytical bounds for quantities of interest in the analysis using a multivariate normal distribution as a null hypothesis, and we show how deviations from normality can be dealt with in a consistent manner.

## 2 Methods

### 2.1 Experimental setting

Four rats were implanted with 6 tetrodes (McNaughton et al. 1983) in the prelimbic and infralimbic areas in the medial PreFrontal Cortex (mPFC). 1692 cells were recorded in the mPFC from four rats, during a total of 63 recording sessions (Rat 15: 16; Rat 18: 11; Rat 19: 12; Rat 20: 24). Rats performed a rule-shift task on a Y maze, where, in order to earn a food reward, they had to select one of the two target arms, on the basis of four rules presented successively. The rules concerned either the arm location, or whether the arm was illuminated (changing randomly at each trial). As soon as the rat learned a rule, the rule was changed and had to be inferred by trial and error. Recordings were made also in sleep periods prior to (PRE) and after (POST) training sessions. For an extensive description of the behavioral and experimental methods, see Peyrache et al. (2009).

### 2.2 Analysis framework

- 1.
Spike trains from multiple, simultaneously recorded cells from the awake epoch are binned and z-transformed.

- 2.
The correlation matrix of the binned spike trains is computed, and diagonalized.

- 3.
The eigenvectors associated to the largest eigenvalues are retained; a threshold value can be computed from the upper bound for eigenvalues of correlation matrices of independent, normally distributed spike trains.

- 4.
Spike trains from the sleep epoch are binned and z-transformed.

- 5.
A measure of the instantaneous similarity (termed

*reactivation strength*) of the sleep multi-unit activity at each time (the*population vector*) with the eigenvalue is computed.

Reactivation strength is a time series describing how much the sleep ensemble activity resembles the awake activity at any given time. To make the claim that *replay* of experience-related patterns is taking place during sleep, we need to test the computed reactivation strength against an appropriate null hypotheses. The simplest null hypothesis is that the sleep activity is completely independent from the awake data, and, for example, is drawn from a multivariate normal distribution. Clearly enough, disproving this hypothesis does not provide sufficient evidence for replay: certain *structural* activity correlations may have existed prior to the experience, perhaps because of already present synaptic connections. Moreover, activity distributions may be non-normal, if for no other reason, because binned spike trains for small enough bin sizes will tend to be very *sparse* and thus much of the mass of the distribution will concentrate around zero, causing the distribution to be strongly asymmetrical. Nevertheless, from the conceptual point of view, this null hypothesis is interesting as it allows to better characterize deviations from normality in the activity distribution and their consequences.

To control for structural correlations, the sleep data must be compared with another sleep session recorded prior to the experience: if reactivation strengths in sleep after experience (the POST epoch) tend to be larger than the values for the same neuron ensemble measured for the sleep before (or the PRE epoch), one may conclude that the experience epoch contributed to increase sleep activity correlations, and replay has taken place. If no difference between PRE and POST reactivation strength is measured, one may conclude that experience had no effect on correlations in spontaneous activity during sleep.

An important feature of this technique is that it permits instantaneous assessments of replay. Formally, reactivation strength measures similarity between the correlation matrices for the awake and sleep data (the approach followed in works such as e.g. Kudrimoti et al. 1999), which was decomposed into the contributions coming from each eigenvector and each time during sleep. As discussed below, this considerably increases analysis power, as replay time series can be correlated with other physiological time series of relevance.

Furthermore, it is also useful to apply the analysis in the opposite sense: extracting templates from replay events in the sleep epoch, and matching them to the awake data. In this way, one can identify those behavioral events with activity most similar to sleep activity, and therefore, which behavioral events may contribute the most to replay. For this reason, we will describe the procedure in terms of generic TEMPLATE and MATCH epochs.

### 2.3 Isolation of neural patterns

Let us consider N neurons recorded simultaneously over the time interval [ 0 ...*T* ]. The neurons’ activity could be represented by a series of spike times noted \(\left\{ t_{j}^{i} \right\}\) where \( t_{j}^{i} \) is the *j*-th spike of the *i*-th neuron (i = 1 ... N).

*B*is the total number of bins and

*t*

_{ k }represents the center time of the bins:

*b*is the bin width (

*b*=

*T*/

*B*). Hereafter, for the sake of clarity, the indices of the discrete times

*t*

_{ k }will be omitted. Then, these binned activities are z-transformed, obtaining the

**Q**matrix:

*pairwise cell activity correlation matrix*is then obtained as

*C*

_{ ij }, are the Pearson correlation coefficients between the spike count series for neurons

*i*and

*j*. To disambiguate the contribution of each pattern in the resulting correlation coefficients, we perform a PCA on the

**Q**matrix, that is, an eigenvector decomposition of the correlation matrix. This yields a set of N eigenvectors \(p^{l}\text{, }l=1..N\), each associated to an eigenvalue λ

_{ l }. The patterns will be associated to the

*projectors*of the correlation matrix, noted

**P**

^{(l)}, which are the outer products of all eigenvectors with themselves, providing the following representation of the correlation matrix:

_{ l }. PCA allows to distinguish these patterns which can, in turn, be compared with the instantaneous cell activity during different epochs. In order to do that though, we also need to establish which patterns are likely to reflect underlying information encoding processes and which are the result of noise fluctuations. This problem is addressed below.

### 2.4 Time course of template matching

Let us consider two epochs, *TEMPLATE* and *MATCH*. The general idea is to compare the instantaneous co-activations of neurons during the MATCH epoch with the patterns identified during TEMPLATE, following the method proposed above.

*MA*−

*TE*stands for

*MATCH-TEMPLATE*. This measure is strongly positive in the case of high similarity and is strongly negative in the case where correlations change sign (from positive to negative and

*vice versa*) from the TEMPLATE to the MATCH epoch.

*M*

^{MA − TE}can be expressed as a sum over a time series defined for each time bin during the POST (or PRE) epoch (by using Eq. (3)):

*B*

^{ MATCH }is the number of bins in the MATCH epoch, and

*t*during the rest epochs, to produce the time series \(R^{MA-TE}_{0}(t)\). \(R^{MA-TE}_{0}(t)\) represents a decomposition of the epoch-wide correlation similarity into its instantaneous contributions, i.e. the similarity between the current population vector at time

*t*and the general pattern of co-activation during the TEMPLATE epoch. Therefore, it contains information on exactly when during the MATCH epochs occur patterns of co-activation similar to those that occurred in TEMPLATE. However, this measure can still combine together factors from several different patterns which may co-activate independently. The obtained time course may therefore be the result of averaging over these distinct patterns, which may in fact behave quite differently from one another. The neural patterns are extracted from

**C**

^{ TEMPLATE }:

*l*-th co-activation template and the ongoing activity. The exclusion of the diagonal terms in Eq. (13) reduces the sensitivity of the reactivation strength measure to fluctuations in the instantaneous firing rates. The mean reactivation measure,

*M*

^{MATCH − TEMPLATE}is therefore a weighted sum of the time-averaged value of pattern similarity:

## 3 Results

### 3.1 Significance of principal components

*B*→ ∞ and

*N*→ ∞, with

*q*=

*B*/

*N*≥ 1 fixed,

*σ*

^{2}is the variance of the elements of the random matrix, which in our case is 1, because the

**Q**matrix is z-transformed. Equation (15) shows that the distribution vanishes for λ greater than an upper limit λ

_{ max }. Under the null hypothesis of a random matrix

**Q**, the correlations between spike trains are determined only by random fluctuations, and the eigenvalues of

**C**must lie between λ

_{ min }and λ

_{ max }. Eigenvalues greater than λ

_{ max }are therefore a sign of non-random effects in the matrix, and for this reason we call principal components associated to those eigenvalues

*signal*components, while those associated to eigenvalues between λ

_{ min }and λ

_{ max }are defined as

*non-signal*components. However, the finite size of data sets implies that eigenvalue distribution borders are not as sharp as the theoretical bounds described by Eq. (15). The highest eigenvalue of any correlation matrix is drawn from the Tracy-Widom distribution (Tracy and Widom 1994) in the case of normal, or close to normal, variables. Thus, the highest eigenvalue lies around λ

_{ max }with a standard deviation of order

*N*

^{ − 2/3}(which assumes a value of ~0.1 for

*N*= 30, typical for our recordings). We also use the value λ

_{ max }to normalize uniformly eigenvalues across different section, defining the normalized encoding strength as

_{ max }) hence allowing rejection the null hypothesis of independent spike count series. To check whether any other irregularities (i.e. normality violation) in the distribution of binned spike trains could affect the eigenvalues of the correlation matrix (see for example Biroli et al. 2007), each row of

**Q**was randomly permuted. The resulting shuffled matrix is composed of rows whose individual distributions are preserved but whose temporal interactions are lost. The spectra of their correlation matrix are shown in Fig. 1(b). All eigenvalues remain within the bounds of the Marc̆enko-Pastur distribution (red curve, equivalent to the ones presented in Fig. 1(a),

*right*). Thus, we argue that the observed signal eigenvalues are an effect of the correlation between spike trains, and not simply an effect of the non-normality of each cell’s binned spike count.

### 3.2 Average reactivation

*t*, the standardized population vector is written \(Q_t = [Q_{1t},\ldots, Q_{it},\ldots, Q_{Nt}]^T\). Let

**W**

_{ t }be defined as \(\mathbf{W}_t = Q_t^TQ_t\).

**W**

_{ t }can be decomposed in a diagonal matrix \(\mathbf{W}^D_t\) and the remaining matrix \(\mathbf{W}^R_t\), therefore:

*POST*epoch, taking the role of the MATCH epoch). In Fig. 2 the epoch-wide time course of

*R*for an example principal component from a recording of an ensemble of mPFC neurons is displayed for the PRE and POST epochs. Transient peaks can be observed that are much stronger in POST, and concentrated in the periods of identified slow wave sleep (SWS) (Peyrache et al. 2009). However, the baseline level is comparable between PRE and POST epochs. For this reason, from this point on, sleep epochs will always refer to SWS only. The SWS preceding the AWAKE epoch is taken as a control (

*PRE*epoch). The variable \(\gamma^{PRE}_l\) (resp. \(\gamma^{POST}_l\)) quantifies the amount of variance that a given eigenvector from AWAKE could explain during the PRE epoch (resp. POST). The empirical distribution of \(\gamma^{PRE}_l\) (resp. \(\gamma^{POST}_l\)) as a function of λ

_{ l }is shown in Fig. 3(a). During POST,

*γ*

_{ l }was more correlated to λ

_{ l }than in PRE, indicating that the correlation structure is more similar to that measured during AWAKE than it is in PRE. Note that, if it held that

**C**

^{ POST }=

**C**

^{ AWAKE }, then \(\gamma^{POST}_l = \uplambda_l\) for all

*l*. In general, this is not the case, for example, because the sleep correlation structure includes patterns that are characteristic of that behavioral phase, and these do not appear during the AWAKE epoch. In any event, during POST the regression line between λ

_{ l }and the corresponding values of

*γ*

_{ l }has a steeper (and closer to 1) slope, indicating that the POST correlation matrix is more similar to its AWAKE analog than the one computed for PRE. Figure 3(b) shows an eigenvector-by-eigenvector (combined across sessions) comparison of the reactivation strengths during PRE and POST. While it is apparent that some reactivation strengths appear during PRE as well, most eigenvectors showed a larger value for POST, especially for eigenvectors associated to large eigenvalues. During PRE, reactivation strengths were nevertheless still important. This could be due, as mentioned above, to structural correlations, as well as to neural processes reflecting anticipation of the upcoming task (or perhaps lingering reactivations of yet earlier experiences).

### 3.3 Distribution of R

In exploring the time course of the reactivation measure *R*, one interesting question that emerges is the nature of its variability. One possibility is that *R* fluctuates steadily around an average value (possibly different for each epoch), as would be the case, for example, if the underlying spike trains were a gaussian process. Alternatively, power-law behavior for the distribution of *R* values would indicate that the temporal evolution of *R* is dominated by strong transients, as it would result, for example, from “avalanche” dynamics (Beggs and Plenz 2003, 2004; Levina et al. 2007). In fact, if spike trains are multivariate normally distributed variables, the distribution of *R* can be computed and compared with experimental data. Let us consider the case in which the TEMPLATE activity is considered fixed, and we shall compute the distribution of *R* when the columns of the *Q* ^{ MATCH } matrix are drawn from a multivariate normal distribution with covariance matrix **C**, \(Q^{MATCH}_{t} \sim \mathcal{N}(0,\mathbf{C})\).

*m*different time bins,

**Q**is a

*m*×

*N*matrix and

**W**=

**Q**

^{ T }

**Q**is a

*N*×

*N*matrix drawn from the so-called Wishart distribution with

*m*degrees of freedom, \(\mathbf{W} \sim \mathbf{W}_N \left (\mathbf{C},m \right )\). It can be shown that, for any given N-dimensional vector

*z*:

*z*=

*p*

^{ l }, and

*C*=

*C*

^{ TEMPLATE }it leads to:

*Q*

_{ it }are drawn from a multivariate normal distribution: where

**C**is the covariance matrix of the multivariate distribution (as the columns of Q are by definition z-transformed,

**C**is also the correlation matrix). From Eq. (13),

*R*

_{ l }could be written as:

**D**is a diagonal matrix whose elements are \(\left\{ \left ( p^l_i \right )^2 \right\}_{i=1..N}\) The two terms on the right side of Eq. (23) should be considered separately: \(\alpha(t) = Q_t^T \mathbf{P}^{(l)} Q_t\) and \(\beta(t) = Q_{t}^T \mathbf{D}^{(l)} Q_{t}\). First,

*α*(

*t*) is easily deduced from Eq. (22)

*C*=

*C*

^{ TEMPLATE }.

*β*(

*t*) is a weighted sum of

*χ*

^{2}distributed variables whose number of degrees of freedom is not known

*a priori*:

*k*and scale parameter

*θ*, this gives (with the superscript of

*p*

^{ l }omitted):

*β*is equivalent to

^{1}

*α*and

*β*are assumed independent, the theoretical distributions of \(R^{TE-TE}_l\) and \(R^{MA-TE}_l\) are:

*α*and

*β*were correlated, the tail of the distribution could not be “heavier” or more skewed than an exponential distribution. Nevertheless, as we shall see in the following section, experimental evidence shows that those distributions are actually power-laws. The distribution of

*R*

^{TE − TE}for the first eigenvectors of the 3 data sets presented in Fig. 1 are plotted in Fig. 4(a), against the theoretical curve (red) and the result from multivariate normal data simulations (blue). The very same distributions are shown in log-log scales in Fig. 4(b) to highlight the power-law tails of the distributions. The higher the encoding strength (λ/λ

_{ max }), the better the tail is fitted with a power-law (in other words the tail is linear in log-log plots). Figure 4(c) shows the theoretical (under the multivariate normal hypothesis) and empirical distribution for the individual terms

*α*and

*β*.

### 3.4 Reactivation is a rare event

**Q**matrix), does not prevent the instantaneous firing rate from varying considerably, for example because of UP/DOWN states bistability dominating cortical activity during sleep (Steriade 2006). In order to control for this possibility, we computed the reactivation strength from shuffled data where, for each time bin, the identity of the cells was randomly permuted. This shuffling procedure preserves the instantaneous global firing rate (and its fluctuations), but it destroys the patterns of co-activation. In Fig. 6(a), from the same session presented in Fig. 2, the reactivation measure was computed for one principal component while the eigenvector weights (or equivalently, the identities of the cells in the multi-unit spike train ) were shuffled 1000 times. The 99th percentile of the resulting distribution, for each time bin, is shown as the grey curve superimposed upon actual reactivation measure data (black dots). Those points are plotted as a function of the average firing rate (in z-score) which represent the global activation of the cell population. There is a relation between instantaneous global activation and the upper bound of the distribution of shuffled measures (the 99th percentile) which is similar in PRE and POST sleep. Nevertheless, while the actual reactivation measure remained within the expected bounds in PRE sleep (only 1% of the bins exceeded the shuffled measure), the actual reactivation measure largely exceeded this confidence interval in POST sleep (3.4% of the bins were above the 99th percentile). To check whether this could be due to a difference in global population activation, the PRE and POST sleep distribution of average z-score were compared (right inset) and, indeed, showed no difference (Kolmogorov-Smirnov test,

*p*= 0.11).

This difference in the tail of the distribution is very important for the excess reactivation strength in POST with respect to PRE which we take as evidence for memory replay. In the example of Fig. 6(a), 4.5% of the bins from POST exceeded the 99th percentile of the PRE distribution. Hence, the difference in tails of PRE and POST distributions (as in the examples in Fig. 5) resulted in a higher probability for reactivation strength values from POST sleep to exceed the 99th percentile of the PRE reactivation strength distribution than the expected 1% (Fig. 6(b)) in an encoding strength dependent manner: the percentage of “outliers” is significantly above chance for all groups of components and it increases with encoding strength. Whereas average actual reactivation strength differences between PRE and POST (Fig. 6(c), significant for all groups, *p* < 0.05, t-test) show the same profile as the increased number of outliers in POST sleep (Fig. 6(b)), there was no difference in mean of the shuffled reactivation strengths. Furthermore, reactivation strengths for shuffled data were on average one order of magnitude smaller than reactivation strengths computed from actual data.

*P*(

*u*) with an exponential tail, this function will reach an asymptotic value, indicating that large values contribute little to the overall average. Diverging values of \(\langle R \rangle_{-\infty}^{r}\) (e.g. ∝ log(

*r*)), are indicative of a

*P*(

*u*) with a tail decaying with a power law. This function converges asymptotically to a value equal to the difference of the average reactivation strength between POST and PRE (also equal to

*γ*

^{ POST }−

*γ*

^{ PRE }). The black diamonds indicate the 99th percentile of the distribution of POST sleep reactivation distribution. Hence, up to half of the difference (for the highest encoding strength) between POST and PRE sleep average of reactivation strength is due to one percent of the time bins from POST sleep, that is, the bins in which the transient reactivation events took place.

### 3.5 Interactions between different cell assemblies

*R*for the two principal components show a trough for zero-lag in the cross-correlation, showing that the simultaneous activation of the two components was less likely than in the case of uncorrelated time course. This effect was observed for all pairs of principal components compute from the same sessions (Peyrache et al. 2009).

*R*correspond to transient synchronization, or co-firing, of cells with same-signed weights in the principal vector. Nevertheless, each spike train participates differently to any particular reactivation strength, likely depending on its associated weight in the principal component. To quantify the contribution of the

*k*th cell, the reactivation strength \(R^{-k}_{l}\) was computed with ∀

*t*,

*Q*

_{ kt }= 0 , or by removing the terms depending on

*Q*

_{ kt }in Eq. (13):

## 4 Discussion

This study shows that a simple and linear pattern separation method such as PCA can be powerful in the identification and characterization of cell assemblies in brain recordings. This is an important part of the study of the replay phenomenon, where two epochs must be compared, one in which assemblies would be encoded, and another one, in which the same assemblies might be replayed again. By construction, our method is a simple extension of the seminal work of Bruce McNaughton and co-workers (Wilson and McNaughton 1994; Kudrimoti et al. 1999), offering two important new features: first, a detailed time course for replay is obtained, at the scale of the chosen temporal bin (100 ms in the present study). The resulting resolution is much finer than what can be achieved if replay is only measured from the similarity between the epoch-wide correlation matrices. This has important consequences for the study of the physiology of replay. In particular, we have found that replay takes place for the most part in discrete, transient events (see e.g. Figs. 2 and 5), which correspond to the coordinated activation of subgroups of cells. In fact, such transients mostly take place during UP states characteristic of slow-wave sleep. These are periods of elevated, relatively steady activation, when measured at the level of the global neuronal population. However, a very different time course is uncovered when we consider the dynamics of subgroups of cells, defined by co-activations measured during wakefulness: a avalanche-like dynamics (Beggs and Plenz 2003, 2004; Levina et al. 2007), which is embedded in a generally more regular population dynamics. Moreover, a detailed view of temporal evolution of replay has allowed to explore the links between this phenomenon on one hand and hippocampal sharp waves (crucial for hippocampal replay (Kudrimoti et al. 1999)) and UP-DOWN state transitions on the other, showing how replay is an integral part of hippocampal-cortical interactions and sleep physiology (Peyrache et al. 2009).

Second, PCA allows to tease apart the dynamics of different cell assemblies, corresponding to different principal components. Interestingly, distinct subgroups tend to seldom reactivate at the same instant, suggesting that some sort of pattern separation mechanism may take place during sleep. Because the time courses of the different principal components are un- (or anti-)correlated (Fig. 8), separating them allows to reveal details of the temporal evolution which would be otherwise averaged out, for example, the transients discussed above.

This measure also lends itself to rigorous mathematical analysis, making some inroads towards precisely defined null hypotheses to be tested against the experimental results. The known eigenvalue spectrum of correlation matrices from purely random data (Marčenko and Pastur 1967; Sengupta and Mitra 1999) allows determination of which of the principal components in a given data set are likely to carry meaningful information; in the data considered here, up to 5 or 6 PCs can be found in a simultaneous recording of 30–50 neurons (Fig. 1). This could be seen as a generalization in N dimensions of the classical Pearson test for pairwise correlations. The boundaries of the support of the theoretical distribution for the eigenvalues, λ_{ min } and λ_{ max }, can be taken as the critical value for the rejection of the null hypothesis. In the range of parameters corresponding to our practical experimental situation (number of variables, N ~ 50, number of time bins, b ~10^{4} for a bin size of 100 ms) these boundaries are sharp, as demonstrated by the Tracy-Widom estimate of the variance of the distribution of the largest eigenvalue (Tracy and Widom 1994). Thus, an analysis procedure that considers as ‘signal’ the principal components associated to an eigenvalue larger than λ_{ max } is justified from the theoretical point of view. This allows us to identify certain principal components as signal-carrying cell assemblies.

In the next stage of the analysis, the *R* measure is computed, measuring the time course of replay during the PRE and the POST epoch. In principle, replay could be the result of a continuous process, for example one that modified the probability of co-activation of cell pairs, as a consequence of synaptic modification. In this case, one would expect an exponentially tailed distribution for the *R* values. This was indeed verified analytically, under the hypothesis of multivariate, normally distributed data.

Reactivation strengths are greater than chance levels both in PRE and POST sleep. This could be due to structural correlations, pre-encoded in the synaptic matrix. Such correlations would be present in ensemble activity at all times, both in spontaneous and in behaviorally evoked activity, and would not have to encode any task-relevant information. It is also possible that, during PRE sleep, the prefrontal cortex is already engaged in processes anticipatory of the task. This could explain the similarities between the activity in the PRE and AWAKE epochs.

Nevertheless, POST sleep shows a significantly greater degree of replay. This can be observed by empirically comparing the *R* distributions for PRE and POST. Interestingly, most of the difference between PRE and POST is accounted for by the very large data points in the tail of the distribution, so that, for the principal components associated to the largest eigenvalues, up to 50% of the difference is accounted for by only the largest 1% of the points. It seems therefore likely that the large transients in the replay measure are at least in part a consequence of replay. Moreover, it is possible that during experience, synaptic plasticity operates by modifying and strengthening existing cell assemblies (during gradual learning for example), as opposed to creating new ones from scratch. This would also contribute to explain why reactivation strength for the same eigenvectors may be high both in PRE and POST (albeit stronger in POST).

Our method, in its current version, has some limitations. For one, it does not provide a way to directly discount structural correlation patterns present in the PRE epoch from the templates extracted by the AWAKE epoch, which would obviate the need for a comparison of the empirical PRE and POST distributions. Also, it would be important to compute analytical bounds for quantities under null hypotheses less stringent than that of multivariate normal spike trains. Still this technique has already led to scientific results of relevance (Peyrache et al. 2009): as another example of use of this technique, as mentioned above, the sleep epoch can serve as a template for detecting matches in the awake epoch: we extracted patterns from PCA applied to the POST epoch, and matched them on the activity during the AWAKE epoch. This allowed us to assess which behavioral phases of the task were represented the most in the sleep activity (and possibly, be preferentially consolidated). We concluded that this coincided with activity at the “choice point” of the maze, i.e. the fork of the Y-maze where the animal had to commit to a potentially costly choice. Also, the effect depended upon learning: sleep-derived activity patterns were more concentrated at the decision point after the rat acquired the rule governing reward (Peyrache et al. 2009). These initial results provide hope that this method, for its relative simplicity and ease of approach with mathematical tools, may spur further experimental and analytical work.

## Footnotes

- 1.
Note that \(\beta \sim (2m)^{-1} \chi^2_{2m}\) such that

*β*follows a normalized*χ*^{2}distribution whose degree of freedom is \(\left (\sum p^4_i \right )^{-1}\).

## Notes

### Acknowledgements

We thank Dr. A. Aubry for interesting discussions. Supported by Fondation Fyssen (FPB), Fondation pour la Recherche Medicale (AP), EC contracts FP6-IST 027819 (ICEA), FP6-IST-027140 (BACS), and FP6-IST-027017 (NeuroProbes).

### **Open Access**

This article is distributed under the terms of the Creative Commons Attribution Noncommercial License which permits any noncommercial use, distribution, and reproduction in any medium, provided the original author(s) and source are credited.

## References

- Beggs, J. M., & Plenz, D. (2003). Neuronal avalanches in neocortical circuits.
*Journal of Neuroscience, 23*(35), 11167–11177.PubMedGoogle Scholar - Beggs, J. M., & Plenz, D. (2004). Neuronal avalanches are diverse and precise activity patterns that are stable for many hours in cortical slice cultures.
*Journal of Neuroscience, 24*(22), 5216–5229.CrossRefPubMedGoogle Scholar - Benchenane, K., Peyrache, A., Khamassi, M., Battaglia, F. P., & Wiener, S. I. (2008). Coherence of theta rhythm between hippocampus and medial prefrontal cortex modulates prefrontal network activity during learning in rats.
*Soc Neuroscie Abstr*, 690.15.Google Scholar - Biroli, G., Bouchaud, J. P., & Potters, M. (2007). On the top eigenvalue of heavy-tailed random matrices.
*EPL (Europhysics Letters), 78*(1), 10001+.CrossRefGoogle Scholar - Bishop, C. M. (1995).
*Neural networks for pattern recognition*. New York: Oxford University Press.Google Scholar - Bourlard, H., & Kamp, Y. (1988). Auto-association by multilayer perceptrons and singular value decomposition.
*Biological Cybernetics, 59*(4), 291–294.CrossRefPubMedGoogle Scholar - Buzsáki, G. (2004). Large-scale recording of neuronal ensembles.
*Nature Neuroscience, 7*(5), 446–451.CrossRefPubMedGoogle Scholar - Chapin, J. K., & Nicolelis, M. A. (1999). Principal component analysis of neuronal ensemble activity reveals multidimensional somatosensory representations.
*Journal of Neuroscience Methods, 94*(1), 121–140.CrossRefPubMedGoogle Scholar - Euston, D. R., Tatsuno, M., & Mcnaughton, B. L. (2007). Fast-forward playback of recent memory sequences in prefrontal cortex during sleep.
*Science, 318*(5853), 1147–1150.CrossRefPubMedGoogle Scholar - Fujisawa, S., Amarasingham, A., Harrison, M. T. T., & Buzsáki, G. (2008). Behavior-dependent short-term assembly dynamics in the medial prefrontal cortex.
*Nature Neuroscience, 11*, 823–833.CrossRefPubMedGoogle Scholar - Geisler, C., Robbe, D., Zugaro, M., Sirota, A., & Buzsáki, G. (2007). Hippocampal place cell assemblies are speed-controlled oscillators.
*Proceedings of the National Acadademy of Science of the United States of America, 104*(19), 8149–8154.CrossRefGoogle Scholar - Harris, K. D., Csicsvari, J., Hirase, H., Dragoi, G., & Buzsáki, G. (2003). Organization of cell assemblies in the hippocampus.
*Nature, 424*(6948), 552–556.CrossRefPubMedGoogle Scholar - Hebb, D. O. (1949).
*The organization of behavior: A neuropsychological theory*. New York: Wiley.Google Scholar - Ikegaya, Y., Aaron, G., Cossart, R., Aronov, D., Lampl, I., Ferster, D., et al. (2004). Synfire chains and cortical songs: Temporal modules of cortical activity.
*Science, 304*(5670), 559–564.CrossRefPubMedGoogle Scholar - Imhof, J. P. (1961). Computing the distribution of quadratic forms in normal variables.
*Biometrika, 48*(3–4), 419–426.Google Scholar - Ji, D., & Wilson, M. A. (2006). Coordinated memory replay in the visual cortex and hippocampus during sleep.
*Nature Neuroscience, 10*(1), 100–107.CrossRefPubMedGoogle Scholar - Kudrimoti, H. S., Barnes, C. A., & McNaughton, B. L. (1999). Reactivation of hippocampal cell assemblies: Effects of behavioral state, experience, and eeg dynamics.
*Journal of Neuroscience, 19*(10), 4090–4101.PubMedGoogle Scholar - Lee, A. K., & Wilson, M. A. (2002). Memory of sequential experience in the hippocampus during slow wave sleep.
*Neuron, 36*(6), 1183–1194.CrossRefPubMedGoogle Scholar - Levina, A., Herrmann, J. M., & Geisel, T. (2007). Dynamical synapses causing self-organized criticality in neural networks
*Nature Physics, 3*(12), 857–860.CrossRefGoogle Scholar - Louie, K., & Wilson, M. A. (2001). Temporally structured replay of awake hippocampal ensemble activity during rapid eye movement sleep.
*Neuron, 29*(1), 145–156.CrossRefPubMedGoogle Scholar - Marr, D. (1971). Simple memory: A theory for archicortex.
*Philosophical transactions of the Royal Society of London. Series B, Biological Sciences, 262*(841), 23–81.CrossRefPubMedGoogle Scholar - Marčenko, V. A., & Pastur, L. A. (1967). Distribution of eigenvalues for some sets of random matrices.
*Mathematics of the USSR-Sbornik, 1*(4), 457–483.CrossRefGoogle Scholar - McNaughton, B. L., O’Keefe, J., & Barnes, C. A. (1983). The stereotrode: A new technique for simultaneous isolation of several single units in the central nervous system from multiple unit records.
*Journal of Neuroscience Methods, 8*(4), 391–397.CrossRefPubMedGoogle Scholar - Mokeichev, A., Okun, M., Barak, O., Katz, Y., Ben-Shahar, O., & Lampl, I. (2007). Stochastic emergence of repeating cortical motifs in spontaneous membrane potential fluctuations
*in vivo*.*Neuron, 53*(3), 413–425.CrossRefPubMedGoogle Scholar - Nádasdy, Z., Hirase, H., Czurkó, A., Csicsvari, J., & Buzsáki, G. (1999). Replay and time compression of recurring spike sequences in the hippocampus.
*Journal of Neuroscience, 19*(21), 9497–9507.PubMedGoogle Scholar - Nadel, L., & Moscovitch, M. (1997). Memory consolidation, retrograde amnesia and the hippocampal complex.
*Current Opinion in Neurobiology, 7*(2), 217–227.CrossRefPubMedGoogle Scholar - Nicolelis, M. A., Baccala, L. A., Lin, R. C., & Chapin, J. K. (1995). Sensorimotor encoding by synchronous neural ensemble activity at multiple levels of the somatosensory system.
*Science, 268*(5215), 1353–1358.CrossRefPubMedGoogle Scholar - Oja, E. (1982). A simplified neuron model as a principal component analyzer.
*Journal of Mathematical Biology, 15*(3), 267–273.CrossRefPubMedGoogle Scholar - Pipa, G., Wheeler, D. W., Singer, W., & Nikolić, D. (2008). Neuroxidence: Reliable and efficient analysis of an excess or deficiency of joint-spike events.
*Journal of Computational Neuroscience, 25*(1), 64–88.CrossRefPubMedGoogle Scholar - Peyrache, A., Khamassi, M., Benchenane, K., Wiener, S. I., & Battaglia, F. P. (2009). Replay of rule learning-related patterns in the prefrontal cortex during sleep.
*Nature Neuroscience*(Advanced Online Publication). doi: 10.1038/nn.2337.PubMedGoogle Scholar - Ribeiro, S., Gervasoni, D., Soares, E. S., Zhou, Y., Lin, S. C., Pantoja, J., et al. (2004). Long-lasting novelty-induced neuronal reverberation during slow-wave sleep in multiple forebrain areas.
*PLoS Biology, 2*(1), e24.CrossRefGoogle Scholar - Riehle, A., Grun, S., Diesmann, M., & Aertsen, A. (1997). Spike synchronization and rate modulation differentially involved in motor cortical function.
*Science, 278*(5345), 1950–1953.CrossRefPubMedGoogle Scholar - Scoville, W. B., & Milner, B. (1957). Loss of recent memory after bilateral hippocampal lesions.
*Journal of Neurology, Neurosurgery and Psychiatry, 20*(11), 11–21.CrossRefGoogle Scholar - Sengupta, A. M., & Mitra, P. P. (1999). Distributions of singular values for some random matrices.
*Physical Review E, 60*(3), 3389+.CrossRefGoogle Scholar - Squire, L. R., & Zola-Morgan, S. (1991). The medial temporal lobe memory system.
*Science, 253*(5026), 1380–1386.CrossRefPubMedGoogle Scholar - Steriade, M. (2006). Grouping of brain rhythms in corticothalamic systems.
*Neuroscience, 137*, 1087–1106.CrossRefPubMedGoogle Scholar - Tracy, C., & Widom, H. (1994). Level-spacing distributions and the airy kernel.
*Communications in Mathematical Physics, 159*(1), 151–174.CrossRefGoogle Scholar - Wigner, E. P. (1955). Characteristic vectors of bordered matrices with infinite dimensions.
*The Annals of Mathematics, 62*(3), 548–564.CrossRefGoogle Scholar - Wilson, M. A., & McNaughton, B. L. (1994). Reactivation of hippocampal ensemble memories during sleep.
*Science, 265*(5172), 676–679.CrossRefPubMedGoogle Scholar - Wilson, R. I., & Laurent, G. (2005). Role of gabaergic inhibition in shaping odor-evoked spatiotemporal patterns in the drosophila antennal lobe.
*The Journal of Neuroscience, 25*(40), 9069–9079.CrossRefPubMedGoogle Scholar