Introduction

Synaptic tagging and capture (STC) provides a well-established hypothesis of the processes underlying synaptic consolidation, which has been investigated by a multitude of experimental and theoretical studies [1,2,3,4,5,6,7,8,9]. However, the dynamics resulting from the interplay between STC and recurrent network structures, which are characteristic for brain structures like the hippocampus and the neocortex [10, 11], and their relation to long-term memory still remain unclear.

Neuronal activity causes the increase or decrease in the transmission strength of synapses via long-term potentiation (LTP) and long-term depression (LTD). The early phase of these phenomena typically lasts for up to a few hours [12, 13]. To be maintained, the synaptic changes have to undergo a consolidation process, which is described by the STC hypothesis [1, 4]. When a group of neurons in a recurrent neural network experiences strong (learning) stimulation, this causes the formation of a Hebbian cell assembly, which thence represents a memory of the stimulation. Hebbian cell assemblies are groups of neurons exhibiting particularly strong synaptic connections, giving rise to facilitated activation of the assembly neurons [14,15,16,17]. In our previous study [9], we developed a computational model and showed that a memory representation can be encoded by the early-phase of long-term plasticity, while STC subsequently transfers the synaptic changes to the late phase, yielding a long-term memory representation. Here, we develop this model further to investigate how the encoding and consolidation of multiple memory representations guides their interactions. This enables us to explore cognitive implications of our two-phase plasticity model.

In the brain, spontaneous activity is ubiquitous and has been shown to inherently arise from deterministic dynamics in neural networks [18, 19]. In theoretical models, spontaneous activity is usually elicited by background noise. This approach is biologically well-grounded through matching power spectra of in vivo neuronal activity on the one hand and of the stochastic processes used to model background noise on the other hand [20, 21]. Spontaneous activity causes the reactivation of existent cell assemblies via neuronal avalanches [22, 23] and thereby becomes relevant for cognition. In the current study, we use the concept of cell assembly reactivation by spontaneous activity to model the reactivation dynamics of multiple long-term memories, learned and consolidated with different protocols, as considered in neuropsychological experiments. Namely, we compare our modeling results to two different classes of experiments from cognitive psychology: free recall experiments, and priming experiments.

Free recall is a memory test that requires subjects to reproduce information from memory, for example a sequence of words, without cuing. While several models have been developed to describe free recall (e.g., [24,25,26,27,28,29,30]), its biological underpinnings remain elusive. With respect to the serial position of the input items, free recall experiments typically exhibit three prominent characteristics [31,32,33,34,35]: 1. primacy, the better recall of the items learned first, 2. contiguity, the increased probability of recalling neighboring items, and 3. recency, the better recall of the most recently learned items. Primacy is commonly related to extended rehearsal, in which the recency effect could play an indirect role [27, 36, 37]. Recency was first thought to be a solely short-term memory-dependent effect, but has then also been found to persist in long-term memory [27, 33, 38,39,40]. The recency effect is usually much stronger than the primacy effect [32, 33, 40], and the two effects are often jointly referred to as serial-position effect. The serial-position effect has been found to be expressed differently in experiments with different delay times before recall [33, 39, 41, 42]. Through these diverse findings, free recall experiments have essentially instigated the distinction between short-term and long-term memory storage in the human brain (also see [34]). To theoretically explain serial-position effects in long-term memory, abstract memory models have been developed [27, 37, 39]. Furthermore, two recent studies [28, 43] have shown that, rather than working-memory associations as in previous models [24, 26], longer-lasting associations between items seem to account for the effects of free recall. Prior familiarity with the learned items seems, however, not to play an important role [44]. With our model, we link neuronal and synaptic dynamics including STC to emergent network principles. Thereby, we complement the literature by providing a mechanistic explanation of the recency effect in long-term memory representations that are learned from scratch.

Priming (not to be confused with primacy) describes a large set of psychological phenomena across a multitude of timescales, which can occur either consciously or unconsciously [45,46,47]. All priming phenomena have in common that the recall of a certain memory is enhanced following its reactivation by a “priming stimulus.” Many studies on priming have targeted processes in the intersection between psychology and economics, for instance, biases in deciding to buy a certain product. As an example, if we see an advertisement of some brand of orange juice that we have had before, and soon afterwards go to the supermarket to do the weekly shopping, the likelihood of buying exactly that brand seen in the advertisement will be increased. Here, the advertisement acts as a priming stimulus [45, 48, 49]. For priming on a timescale of seconds, Mongillo et al. [50] have provided a model featuring the interplay of short-term plasticity and long-term plasticity. Furthermore, the authors discussed the possibility of transferring this concept to the processes on the next, longer timescales: early- and late-phase long-term plasticity. We take on this idea and show that the STC mechanisms can account for a priming effect on a timescale of hours.

Methods

Model

To simulate the dynamics of memory representations, we used a network model that comprises spiking neurons and synapses with detailed plasticity features. In this subsection, we first present our mathematical description of neurons and synapses, which is depicted in Fig. 1a, as well as the structure of our network at the population level. After that, we explain the different overlap paradigms and stimulation protocols. The parameters that we used are given in Tables 1 and 2. The model is based on our previous study [9], which should be consulted for a more detailed description.

Fig. 1
figure 1

Model and considered protocols. a, b Schematic of the synaptic and network model (adapted from [9]). The synaptic model integrates the interplay between different mechanisms of calcium- and rate-dependent synaptic plasticity and the STC hypothesis. The neural network consists of excitatory (light and dark blue and green disks) and inhibitory neurons (red disks) and receives external input from other brain areas. Only synapses between excitatory neurons undergo plastic changes via the processes shown in (a). Hebbian cell assemblies representing memories consist of groups of strongly interconnected neurons (dark blue, vivid green). The assemblies in b overlap by a fraction of their neurons (dark green). For more details, see the main text. c Different organizational paradigms are defined by different overlap relationships between the three cell assemblies. Pictures show possible real-life examples (all adapted from pixabay.com). d Protocol of the learning phase: the three memory representations are learned sequentially, with three-seconds breaks in between. e Protocol for the consolidation and recall phase during which consolidation takes place for eight hours after learning. Then, spontaneous activity in the network, driven by background noise, is observed for three minutes. Time spans in d and e are similar to those in free recall experiments [32, 34, 41]

Table 1 Values were used as given in this table, unless stated otherwise
Table 2 Values were used as given in this table, unless stated otherwise

Neuron Model

The dynamics of the membrane potential \(V_i(t)\) of the Leaky Integrate-and-Fire neuron i is described by [51]:

$$\begin{aligned} \tau _\mathrm {m}\frac{dV_i(t)}{dt} =& - V_i(t) + V_\mathrm {rev} \\& + \sum \limits _j\sum \limits _{t^k_j} w_{ji} \cdot \exp \left( -(t-t^k_j -t_\mathrm {ax,delay})/\tau _\mathrm {syn}\right)\\& + R_\mathrm {m}\,\left( I_\mathrm {bg}(t)+I_\mathrm {stim}(t)\right) \end{aligned}$$
(1)

with reversal potential \(V_\mathrm {rev}\), membrane time constant \(\tau _\mathrm {m}\), membrane resistance \(R_\mathrm {m}\), synaptic weights \(w_{ji}\), spike times \(t^k_j\), axonal delay time \(t_\mathrm {ax,delay}\), synaptic time constant \(\tau _\mathrm {syn}\), external background current \(I_\mathrm {bg}(t)\), and external stimulus current \(I_\mathrm {stim}(t)\). If \(V_i\) crosses the threshold \(V_\mathrm {th}\), a spike is generated. The spike time \(t^n_i\) is then stored, and the membrane potential is reset to \(V_\mathrm {reset}\), where it remains for the refractory period \(t_\mathrm {ref}\). Under basal conditions, the membrane potential dynamics is mainly driven by an external background current that accounts for synaptic inputs from outside the network, described by an Ornstein–Uhlenbeck process:

$$\begin{aligned} \tau _\mathrm {syn}\frac{dI_\mathrm {bg}(t)}{dt} = - I_\mathrm {bg}(t) + I_0 + \sigma _\mathrm {wn}\cdot \Gamma (t) \end{aligned}$$
(2)

with mean current \(I_0\) and white-noise standard deviation \(\sigma _\mathrm {wn}\). In this equation, \(\Gamma (t)\) is Gaussian white noise with mean zero and standard deviation \(1/\sqrt{dt}\) that approaches infinity for \(dt \rightarrow 0\). The Ornstein–Uhlenbeck process has the same colored-noise power spectrum as the fluctuating input to cortical neurons coming from a large presynaptic population and is therefore well-suited to model the background noise in our model [20].

In addition to the background noise, we use another Ornstein–Uhlenbeck process to model the stimulus current \(I_\mathrm {stim}(t)\) for learning, recall, and priming (cf. the subsection “Learning, Recall, and Priming Stimulation” below):

$$\begin{aligned} \tau _\mathrm {syn}\frac{dI_\mathrm {stim}(t)}{dt} =& - I_\mathrm {stim}(t) \\& + \left( N_\mathrm {stim}\cdot f_\mathrm {stim} + \sqrt{ N_\mathrm {stim}\cdot f_\mathrm {stim}} \cdot \Gamma (t)\right) \cdot 1\,\mathrm {s}\cdot \frac{h_0}{R_\mathrm {m}}. \end{aligned}$$
(3)

Mean and standard deviation of this process are defined by putative input spikes from \(N_\mathrm {stim}\) neurons, occurring at the frequency \(f_\mathrm {stim}\) and conveyed by synapses of weight \(h_0\) (for the unit time of one second) to the neurons of our network that are to be stimulated.

Model of Synapses and STC

All spikes k that occur in neuron j are transmitted to i, if there is a synaptic connection from neuron j to neuron i. The postsynaptic current caused by a presynaptic spike then depends on the total weight or strength of the synapse, which is given by:

$$\begin{aligned} w_{ji} = {\left\{ \begin{array}{ll} \quad h_{ji} + h_0 \cdot z_{ji} & \mathrm {for \ E}\rightarrow \mathrm {E}, \\ \quad w_\mathrm {ei} & \mathrm {for \ E} \rightarrow \mathrm {I}, \\ \quad w_\mathrm {ie} & \mathrm {for \ I}\rightarrow \mathrm {E}, \\ \quad w_\mathrm {ii} &\mathrm {for \ I}\rightarrow \mathrm {I}. \end{array}\right. } \end{aligned}$$
(4)

Based on our previous study [9], we selected the values of the inhibition strength \(w_\mathrm {ie}\) and \(w_\mathrm {ii}\) (Table 1) such that we obtained excitation–inhibition balance with the same background current (cf. Eq. 2) as in [9].

While in the model all synaptic connections involving inhibitory neurons are constant, the weight of E\(\rightarrow\)E connections comprises two variable contributions to account for STC: the early-phase weight \(h_{ji}\), and the late-phase weight \(z_{ji}\). The dynamics of the early-phase weight is given by

$$\begin{aligned} \tau _h \frac{dh_{ji}(t)}{dt} =& \ 0.1\,(h_0 - h_{ji}(t)) \\ & + b_{ji}(t) \cdot \big[ \gamma _\mathrm {p}\cdot (10\,\mathrm {mV}-h_{ji}(t))\cdot \Theta (c_{ji}(t)-\theta _\mathrm {p}) -\gamma _\mathrm {d}\cdot h_{ji}(t)\cdot \Theta (c_{ji}(t)-\theta _\mathrm {d}) + \xi (t)\big], \end{aligned}$$
(5)

where \(\Theta (\cdot )\) is the Heaviside function, \(\tau _h\) is a time constant, \(c_{ji}(t)\) is the calcium concentration at the postsynaptic site, and \(b_{ji}(t)\) describes additional special conditions for the induction of plasticity. The first term on the right-hand side describes the decay of the early-phase weight to the median initial value \(h_0\), the second term describes early-phase LTP with rate \(\gamma _\mathrm {p}\), given that calcium is above the threshold \(\theta _\mathrm {p}\), the third term describes early-phase LTD with rate \(\gamma _\mathrm {d}\), given that calcium is above the threshold \(\theta _\mathrm {d}\), and the fourth term \(\xi (t) = \sqrt{\tau _h \left[ \Theta \left( c_{ji}(t)-\theta _\mathrm {p}\right) + \Theta \left( c_{ji}(t)-\theta _\mathrm {d}\right) \right] }\,\sigma _\mathrm {pl}\,\Gamma (t)\) describes calcium-dependent fluctuations with standard deviation \(\sigma _\mathrm {pl}\) and Gaussian white noise \(\Gamma (t)\) with mean zero and variance 1/dt. In the case of a no-LTD protocol, we set \(\gamma _\mathrm {d}\equiv 0\) and \(\xi (t) = \sqrt{2 \tau _h \Theta \left( c_{ji}(t)-\theta _\mathrm {p}\right) }\,\sigma _\mathrm {pl}\,\Gamma (t)\).

The calcium concentration \(c_{ji}(t)\) at the postsynaptic site is driven by the pre- and postsynaptic spikes n and m, respectively:

$$\begin{aligned} \frac{dc_{ji}(t)}{dt} = -\frac{c_{ji}(t)}{\tau _\mathrm {c}} + c_\mathrm {pre}\,\sum \limits _{n}\delta (t-t^n_j-t_\mathrm {c,delay}) + c_\mathrm {post}\,\sum \limits _{m}\delta (t-t^m_i), \end{aligned}$$
(6)

where \(\delta (\cdot )\) is the Dirac delta distribution, \(\tau _c\) is a time constant, \(c_\mathrm {pre}\) is the contribution of presynaptic spikes, \(c_\mathrm {post}\) is the contribution of postsynaptic spikes, \(t^n_j\) and \(t^m_i\) are spike times, and \(t_\mathrm {c,delay}\) is the delay of calcium triggered by presynaptic spikes.

In addition to the calcium dependence of early-phase plasticity, which is based on previous studies [9, 52,53,54], we imposed special conditions for the occurrence of plasticity to enable the stable formation, consolidation, and activation of large cell assemblies. In their original study [52], Graupner and Brunel had considered pre- and postsynaptic firing rates up to \(\mathrm {50\,Hz}\). However, a drawback in their calcium-based model is the chance that high spike rates on only one side of the synapse suffice to trigger LTP, which is physiologically rather not plausible. To account for the evidence that the firing of only one neuron (i.e., on one side of the synapse) is not sufficient to evoke LTP [12, 14, 55], we introduced an additional threshold \(\nu _\mathrm {th}\) which allows the induction of LTP only if both pre- and postsynaptic firing rate cross this threshold. Furthermore, we blocked early-phase plasticity outside the learning and priming phases to enable fast computation and stable activation dynamics. This is explained by a lack of novelty outside the learning and priming phases, which goes along with lowered neuromodulator concentrations [56, 57] and thereby prevents learning, as experimental studies have demonstrated [58, 59]. In summary, the two “special plasticity conditions” that we added to the model (cf. Fig. 1a) can be put down formally in the following way:

$$\begin{aligned} b_{ji}(t) = & \ \big[ \Theta (c_{ji}(t)-\theta _\mathrm {p})\cdot \Theta (\nu _i(t) - \nu _\mathrm {th}) \cdot \Theta (\nu _j(t) - \nu _\mathrm {th}) \\& + \Theta (c_{ji}(t)-\theta _\mathrm {d})\cdot \Theta (\theta _\mathrm {p}-c_{ji}(t))\big] \cdot NM , \end{aligned}$$
(7)

where the first term in the square brackets describes the LTP induction depending on the calcium concentration \(c_{ji}\) and the firing rate threshold \(\nu _\mathrm {th}\), and the second term describes the LTD induction which only depends on calcium. The firing rate function \(\nu _i\left( t\right)\) describes the firing rate of neuron i at time t, computed using a sliding window of \(0.5\,\mathrm {s}\). The factor \(NM\) is the neuromodulatory impact, which we assume unity during learning and priming phases, and zero otherwise.

Driven by the calcium-based early-phase dynamics, the late-phase synaptic weight is given by:

$$\begin{aligned} \tau _{z}\frac{dz_{ji}(t)}{dt} =& \ p_i(t) \cdot (1-z_{ji}(t))\cdot \Theta ((h_{ji}(t)-h_0)-\theta _\mathrm {tag})\\ &- p_i(t) \cdot (z_{ji} + 0.5)\cdot \Theta ((h_0-h_{ji}(t))-\theta _\mathrm {tag}). \end{aligned}$$
(8)

with the protein amount \(p_i(t)\), the late-phase time constant \(\tau _z\), and the tagging threshold \(\theta _\mathrm {tag}\). The first term on the right-hand side describes late-phase LTP, while the second one describes late-phase LTD. Both depend on the amount of proteins being available. If the early-phase weight change \(\left| h_{ji}(t)-h_0\right|\) exceeds the tagging threshold, the synapse is tagged. This can be the case either for positive or for negative weight changes. The presence of the tag leads to the capture of proteins (if \(p_i(t) > 0\)) and thereby gives rise to changes in the late-phase weight.

The synthesis of new proteins requires sufficient changes in early-phase weights, while there is also an inherent decay of the protein amount [5]:

$$\begin{aligned} \tau _p \frac{dp_i(t)}{dt} = -p_i(t) + \alpha \, \Theta \left[ \left( \sum \limits _j \left| h_{ji}(t)-h_0\right| \right) -\theta _\mathrm {pro}\right] . \end{aligned}$$
(9)

For the major part of this study, early-phase weights were initialized with the value \(h_0\), while late-phase weights were initialized with the value 0. However, for the results shown in Suppl. Fig. S2, the weights were initialized according to a log-normal distribution, which is typical for synapses in the brain [60, 61]. In that case, the early-phase weights were initialized following the distribution:

$$\begin{aligned} h_\text {init} = h_0 \cdot \exp \left( \sigma _h \cdot \mathcal {N}(0, 1)\right) , \end{aligned}$$
(10)

with the standard normal variable \(\mathcal {N}(0, 1)\) and the standard deviation \(\sigma _h=h_0/3\). The initial late-phase weights used for the results in Suppl. Fig. S2 were drawn according to the following distribution:

$$\begin{aligned} z_\text {init} = 0.5 \cdot \exp \left( \sigma _z \cdot \mathcal {N}(0, 1)\right) -0.5, \end{aligned}$$
(11)

with the standard deviation \(\sigma _z=1/3\).

Population Structure

Using the neuron model and the synapse model that have been explained above, we set up a neural network consisting of 2500 excitatory and 625 inhibitory neurons (depicted in Fig. 1b). The ratio of 4:1 between excitatory and inhibitory neurons is typical for cortical and hippocampal networks [62]. Some of the excitatory neurons receive specific inputs to form cell assemblies (see the subsection “Learning, Recall, and Priming Stimulation”). The inhibitory population provides feedback inhibition, while not directly being subject to plastic changes. The probability for the existence of a synapse between two neurons is 0.1, across the whole network. This is a plausible value for diverse cortical areas as well as for hippocampal region CA3 [11, 63]. For every trial, a new network structure is drawn.

Learning, Recall, and Priming Stimulation

Before we apply the first learning stimulus to our network, we let the initial activity settle for 10.0 seconds. Then, we apply our learning protocol, which delivers three stimulus pulses, of 0.1 seconds duration each, to the neurons dedicated to the first cell assembly (A, the first 600 neurons in the network). Breaks of 0.4 seconds separate the pulses. The stimulation is delivered by an Ornstein–Uhlenbeck process (see Eq. 3 introduced above). After the last pulse has ended, we let the network run under basal conditions for 3.0 seconds (Fig. 1d). For the case of intermediate consolidation, which we consider in one part of our study, we extend this period to 28800.0 seconds, such that the synaptic weights may undergo consolidation through synaptic tagging and capture. Every time that stimulation ends, we wait for 12.5 seconds and then disable early-phase plasticity (except for the decay of early-phase changes; also cf. Eqs. 5 and 7). This is reasonable considering the impact of novelty and neuromodulation on encoding [58, 59].

Then, we apply another learning stimulus of the same kind as the first one to the neurons dedicated to the second cell assembly B. In the non-overlapping case, this set of neurons is completely disjunct from the first set. In the cases with overlaps, 10% of the neurons in the assembly (in the main paper; also cf. Suppl. Table S1) are shared with the other assembly. The size of the overlaps is of the same order of magnitude as experimental findings, which range from around 1% to more than 40% [64,65,66,67]. After the learning stimulus we let the network again run under basal conditions for 3.0 seconds (28800.0 seconds in the intermediate consolidation case).

Finally, we apply a third learning stimulus of the same kind, this time to the neurons dedicated to the third cell assembly C. In the non-overlapping case, this set of neurons is completely disjunct from the first and second set. For the overlap cases, still, 10% of the neurons between each two assemblies are shared. We denote the exclusive intersections between each two assemblies by \(I_ {AB}\), \(I_ {AC}\), \(I_ {BC}\) (see Fig. 1c), which may make up for half the overlap or for the whole overlap between two assemblies. If existent, the intersection between all three assemblies \(I_ {ABC}\) makes up for the other half of each overlap between two assemblies (cf. Suppl. Table S1). Specifically, the different subpopulations are set to occupy the following neuron indices in the network (example for overlapping case/“OVERLAP10”):

  • \(I_ {ABC}\): 540–569,

  • \(I_ {AC}\): 0–29,

  • \(\tilde{A}\): 30–539,

  • \(I_ {AB}\): 570–599,

  • \(\tilde{B}\): 600–1109,

  • \(I_ {BC}\): 1110–1139,

  • \(\tilde{C}\): 1140–1649.

By \(\tilde{A}\), \(\tilde{B}\), and \(\tilde{C}\), we denote the subsets of neurons that exclusively belong to one of the assemblies. Note that in the non-overlapping case, \(\tilde{A}=A\), \(\tilde{B}=B\), and \(\tilde{C}=C\).

At two points in our study, we use interleaved learning of the three memory representations. For this, instead of the three steps A-B-C, we employ the following scheme of 15 learning steps: A-B-C-A-C-B-B-A-C-A-B-C-C-B-A.

After learning, we let the network first run under basal conditions for 3.0 seconds and then consolidate for 28800.0 seconds (eight hours, see Fig. 1e). This causes all early-phase weight changes to either decay or be transferred to the late phase, thereby yielding long-term memory representations. After that, we start measuring the spontaneous activity (see subsection below).

For the testing of memory function, shown in Fig. 2, we apply a recall stimulus to randomly drawn 20% of the neurons of an assembly, delivered by one pulse 0.1 seconds long, with otherwise the same parameters as for the learning stimulus. During this, we leave early-plasticity switched off.

Fig. 2
figure 2

Three non-overlapping memory representations are successfully learned and consolidated. Learning order: A-B-C. a Raster plot showing the spiking in the whole network before and after learning and consolidation. At specified times, additional input is applied to 20% of the neurons of one assembly. Each assembly comprises 600 neurons. Spikes of excitatory and inhibitory neurons are shown by blue and red dots, respectively. Avalanches occurring in an assembly are indicated by bars in the respective color. b Firing rates corresponding to the raster plots (computed via bins of \(10\,\mathrm {ms}\)). Shaded areas indicate the time frame of additional stimulation. The dashed line indicates the detection threshold for avalanches. In the two leftmost panels, gray bars next to the peaks emphasize the increase after consolidation in the mean value of the firing rate of A during the induced activation (\(+26.3\%\)) and in the duration of the activation (\(+11.1\%\)). The measure Q based on the stimulation for learning and recall also indicates pattern completion (cf. [9]). c Weight distribution of the synapses within the assemblies, revealing the spread of LTP and LTD. Weight values were discretized into 100 bins, and the relative frequency was normalized across all bins and all groups of synapses. d Abstract weight matrix showing the mean weight within and between all subpopulations of the excitatory population. The asymmetric nature of the matrix reveals the directionality of the coupling strengths. Data in c and d were averaged over 10 trials and taken after learning and consolidation relative to the initial value before learning (\(h_0\))

For the priming of a memory representation, which we investigate in the last part of our study, we apply a stimulus to all neurons of one selected assembly (alternatively, to randomly drawn 50% of the assembly neurons, the results of which are only shown in the Supplementary Information). The stimulus is delivered via one pulse 0.1 seconds long, with otherwise the same parameters as for the learning stimulus (cf. Fig. 6a). The priming stimulus is followed by a period of consolidation (and early-phase decay) of varied duration, after which we start measuring the spontaneous activity.

Determining Avalanches in the Spontaneous Activity

After learning, consolidation, and possibly priming of the cell assemblies, we measure the network activity across three minutes to evaluate the spontaneous activation of the assemblies. To this end, we compute the likelihood that an avalanche with a certain minimum size occurs in a specific assembly, averaged across the whole period of three minutes. In addition, we measure the plain mean firing rate of each assembly and the control population across the whole period of three minutes.

To determine if an avalanche occurs, we count the number of spikes that are produced by the neurons of each cell assembly in a time frame of \(10\,\mathrm {ms}\). Then, we compare these numbers to the threshold value \(n_\mathrm {thresh}\) and consider an assembly active if the threshold is reached. Thus, multiple assemblies may be active at the same time. Finally, we compute the likelihood of an avalanche in assembly A, for instance, by dividing the number of periods \(N_\mathrm {act}(A)\) in which the assembly was active by the total number of time frames (given by the total duration of the simulation \(t_\mathrm {max}/10\,\mathrm {ms}\)):

$$\begin{aligned} l(A) := \frac{N_\mathrm {act}(A)}{t_\mathrm {max}/10\,\mathrm {ms}}. \end{aligned}$$
(12)

The likelihoods for avalanches in B and C are computed analogously. The threshold value \(n_\mathrm {thresh}\) is chosen to be about the border value of the 99% quantile for the number of spikes of an assembly in one time frame (cf. Suppl. Fig. S11).

Computational Implementation and Software Used

We have used C++ in the ISO 2011 standard to implement our simulations. To compile and link the code, we employed g++ in version 7.5.0 with boost in version 1.65.1.

For further data analysis, we used Python 3.7.3 with NumPy 1.20.1 and pandas 1.0.3. For the creation of plots, we used gnuplot 5.0.3, as well as Matplotlib 3.1.0.

We generated random numbers using the generator ‘minstd_rand0’ from the C++ standard library, while the system time served as the seed. We implemented a loop in our code which ensured that for each distribution a unique seed was used. The simulations that had to be performed for this study were computationally extremely demanding (each running for several hours). Thus, it was very fortunate that we had the opportunity to run most of our jobs on the computing cluster of the Gesellschaft für wissenschaftliche Datenverarbeitung mbH Göttingen (GWDG).

All data underlying the results presented in this study are contained in this manuscript and the Supplementary Information. The data can be reproduced using the simulation code and the analysis scripts that we have released under the Apache-2.0 license [68]. The code continues to be developed on https://github.com/jlubo/memory-consolidation-stc.

Results

In this study, we have employed a spiking recurrent neural network model based on a previous model of memory consolidation via synaptic tagging and capture (STC) [9]. The STC principles provide a biologically plausible underpinning of the long-term maintenance of synaptic plasticity and, thereby, of long-term memory representations [4, 8, 9, 55]. In a proof-of-principle approach, we investigate multiple paradigms of the formation, consolidation, and interaction of three cell assemblies in the recurrent network. In the first part of our study, we measure the likelihood of activation of these three assemblies and show that two factors mainly guide their activation: 1. the learning order, and 2. the type of overlaps between the assemblies. By mechanistically describing the learning order effect, our model provides a possible neural correlate for the recency effect in free recall experiments. In the second part of our study, we focus on the functional consequences of STC with respect to such interactions. To this end, we first show that intermediate consolidation phases in the stimulation protocol strongly alter the learning order effect. Thus, we make predictions about memory dynamics on different time scales which can be tested in neuropsychological experiments. Furthermore, depending on consolidation, we encounter a differential effect of hub-like overlap structures, indicating that STC-based mechanisms might be linked to the diverse behavior observed for similarity of memories, specifically, to retroactive interference and enhanced recall [69,70,71,72,73]. As a final step, we show that the interaction between early- and late-phase long-term plasticity constitutes a synaptic mechanism for priming phenomena lasting from minutes to hours.

The neural network that we used consisted of 2500 excitatory and 625 inhibitory leaky integrate-and-fire neurons. To form a cell assembly, we applied strong stimulation to a subset of 600 excitatory neurons in the network and thereby elicited synaptic plasticity, modeled by a spike-timing-dependent calcium model and a firing rate condition. This early-phase plasticity could be followed by consolidation via STC, which transfers substantial weight changes (i.e., weight changes that have sufficed to set a tag) to the long-lasting late phase. The synaptic and neuronal level of the model is depicted in Fig. 1a, linked to a schematic of the network level in Fig. 1b (for more details, see “Methods” and [9]).

After phases of learning and consolidation, we investigated the activation of the assemblies by considering neuronal avalanches within each assembly [22]. To consider an assembly active, we utilized the 99% quantile of the number of spikes within a time window as threshold value. Hence, we considered the 1% of the time windows with most spikes (for more details, see “Methods” and Suppl. Fig. S11). Note that in contrast to the colloquial sense of the word “avalanche,” neural avalanches are defined to refer, depending on the time window, not only to longer but also to relatively short periods of activation.

Indicated by findings on precisely timed events [74,75,76], the likelihood of occurrence of neuronal avalanches seems more relevant to behavioral function than the mere average firing rate within each assembly. Nevertheless, by comparing the results of the likelihood of avalanche occurrence to the average firing rates (Suppl. Figs. S6 and S7), we found that similar conclusions can be drawn for both measures.

Learning and Consolidation of Multiple Long-term Memory Representations

Before investigating the spontaneous activation of long-term memory representations, we will demonstrate how the network learns and consolidates them. We start from an unbiased network in which all weights have the same value. Then, learning proceeds according to the protocol shown in Fig. 1d, followed by consolidation and the investigation of the spontaneous activity shown in Fig. 1e. Substantially extended breaks between learning stimuli, such that “intermediate consolidation” occurs in those phases, are considered in a later part of this study. In the main text, we consider four organizational paradigms, shown in Fig. 1c. The differences between these paradigms are visualized by small pictures. The different paradigms will be examined in the next subsection, while we restrict ourselves to the non-overlapping case in this subsection.

As a first step, after learning and consolidating three non-overlapping memory representations, we tested whether all three assemblies served as functional memory representations. To this end, we examined the ability to recall the learned and consolidated assemblies by applying stimulation to 20% of the neurons of each assembly, subsequently (Fig. 2a, b). We found the response caused by this stimulation to be stronger and to last longer than the response to the same stimulation in a naive network before learning (the example for A is shown in Fig. 2a,b in the two leftmost panels), which confirms the functionality of the memory representations. Moreover, we observed avalanches within the assemblies during spontaneous activity after learning and consolidation (Fig. 2a, b). Note that the synchronicity in the activity is also increased after learning and consolidation (Fig. 2a). Since the average excitatory synaptic weight in our network after learning and consolidation is increased (\(103.8\%\), integrated over all weights in Fig. 2c), the synchronicity can be explained by findings on network dynamics made by Brunel [77]. These findings show that the strengthening of excitatory connections in a network causes a shift toward a regime of synchronized regular activity. Thereby, we conclude that our network operates in the critical regime close to the phase transition, which is typical for the occurrence of avalanches [22, 23].

In addition to the activity, we scrutinized the weight structure in the network after learning and consolidation. As expected, we found that the weights within each assembly are much higher than the weights in the remainder of the network (Fig. 2c). Remarkably, however, the weight distributions of the three assemblies differ. This is already a hint to the bias in assembly activation introduced by the learning order, which we will investigate in the following subsection. Complementary to the weight distribution, the abstract weight matrix in Fig. 2d reveals the directionality of the coupling strengths within and between subpopulations of the network after learning and consolidation. The matrix reflects the highly potentiated weights within each cell assembly, but also the depressed weights between the assemblies. Presumably, this depression arises from activation of the other assemblies upon learning a new assembly. The weights involving control neurons, on the other hand, exhibit only slight plasticity. Remarkably, synapses incoming from assembly neurons to control neurons are on average slightly depressed, while synapses outgoing from control neurons remain unchanged. Comparing with the weights after learning but before consolidation (Suppl. Fig. S1) reveals that at that time there is also depression in the synapses outgoing from control neurons, which is yet too weak to become consolidated. The depression of the weights presynaptic to the control neurons probably also emerges during learning of the assemblies, while the weights postsynaptic to the control neurons only receive indirect activation during learning. The higher calcium increase through presynaptic spikes in our model might increase this effect further.

Our results on learning, consolidation, and the interaction between memory representations (investigated below) also hold considering randomly initialized weights that follow a log-normal distribution (Suppl. Fig. S2). Such a distribution has been found for synaptic strengths in the brain (cf. [60, 61]) and can therefore mimic the influence of previously acquired memory representations in the network.

For the overlapping paradigm, the weight distribution and the abstract weight matrix are given in Suppl. Figs. S3 and S4. Those results reveal further asymmetries for the weights of the intersections between assemblies, exemplifying the rich dynamics covered by our model.

Learning Order and Overlaps Guide Spontaneous Activity

After the learning and consolidation of the three assemblies, we systematically investigated the spontaneous reactivation of the assemblies driven by background noise. We collected data from three minutes, which is a time span similar to the recall period in free recall experiments [32, 34, 41]. A sample of spontaneous activity and the detected avalanches is shown in Fig. 3a, b.

Fig. 3
figure 3

Spontaneous activity of long-term memory representations, driven by background noise. a Raster plot showing the spontaneous spiking activity in the whole network induced by background noise. Each assembly comprises 600 excitatory neurons (spikes indicated by blue dots). Spikes of inhibitory neurons are shown by red dots. Avalanches occurring in the assemblies are indicated by bars in the assembly colors. Assemblies do not overlap and are learned in the order: A-B-C. b Firing rates corresponding to the raster plot (computed via bins of \(10\,\mathrm {ms}\)). The dashed line indicates the threshold for the detection of an avalanche. c Overview of the likelihood of avalanches in different organizational paradigms, learned with standard protocol (order: A-B-C). Averaged over mean likelihood from 10 networks. Error bars show the 95% confidence interval. Braces and plus/minus signs indicate differences to the “Overlap 10%” paradigm. d Overview of the likelihood of avalanches as in c, but learned with extended, interleaved protocol (order: A-B-C-A-C-B-B-A-C-A-B-C-C-B-A). Data in c and d were averaged over 10 trials

Overlap of memory representations on the network level is related to memory similarity on the cognitive level [28, 29, 64, 66, 78, 79] and is therefore of huge relevance to the understanding of memory dynamics. Here, we consider four different paradigms of overlaps. These are shown in Fig. 1c (also see “Methods”), where small images picture possible real-life examples of non-overlapping and overlapping memory representations: representations of entities from three different categories, like chicken, house, and bicycle, do not overlap, while representations from the same category, like three types of birds, overlap. Mixed cases occur if two entities are in the same category while there is no common category for all three, for instance, bicycle and horse are both means of transportation, and chicken and horse are both farm animals. In the Supplementary Information to our study, we present results on more paradigms, including such with overlaps of varied percentage (Suppl. Figs. S6 and S7). For an overview over these paradigms, see Suppl. Table S1. The most important result from these additional paradigms is that in the majority of cases, larger overlaps further increase the effects that we describe in the following.

Considering the mean likelihood of activation in each of the four paradigms from Fig. 1c with assemblies that were learned in the order A-B-C, we encounter the first main result of our study: the learning order significantly biases the spontaneous activity, such that the activation of more recently learned assemblies is promoted (shown in Fig. 3c). While this is already the case without overlap, overlap between assemblies increases the effect even further.

Our second finding is indicated in Fig. 3c by braces with a plus or minus sign. The braces demonstrate differences between the paradigms with less overlaps and the “Overlap 10%” paradigm. The lack of an overlap between two assemblies (in the abundance of the other overlaps) creates a hub-like structure, where the assembly that overlaps with the two other assemblies is the hub. We find that such hub-like structure reduces the learning order effect for the two assemblies with less overlaps. This means, the activation of the assembly that is disadvantaged by the learning order effect (A in the “no AC” paradigm, B in the “no BC” paradigm) is increased, while the activation of the assembly favored by the learning order effect (C) is decreased. As we will show below, this effect as well as the learning order effect itself are essentially caused by LTD.

For gathering further knowledge about the learning order effect and to examine other effects without the bias of the learning order, we tried to eradicate the learning order effect (also cf. the following subsection on priming). To this end, we employed a learning protocol for the interleaved learning of the three assemblies, using an extended sequence with the order A-B-C-A-C-B-B-A-C-A-B-C-C-B-A (random with the constraint that every assembly occurs equally often). Compared to Fig. 1d, the interleaved protocol simply contains more steps, but otherwise stays the same. In the non-overlapping paradigm, the interleaved learning successfully eradicates any significant impact of the learning order (Fig. 3d). In the other paradigms, however, a certain impact remains. Note that the most recently learned assembly in the extended sequence is A, which is also the assembly reactivated most often.

In our findings shown above, we had seen lower weights in the previously learned assemblies as compared to the newly learned ones (cf. Fig. 2c). Assuming that there might be a decrease in weights (i.e., LTD) in the pre-existing assemblies upon learning a new assembly, we considered the temporal evolution of the synaptic weights with our standard protocol (Fig. 4a). The temporal traces show that the weights within assembly A decrease during learning stimulation for B and C. Comparing these results to the outcome of a standard protocol in which we blocked LTD (Fig. 4c; cf. “Methods”), we see that the decrease is also blocked. A similar difference is evident in the late phase (Fig. 4b, d). We further found that the weight distribution in the case where LTD is blocked is almost equal for all three assemblies (Fig. 5a and Suppl. Fig. S5c), which is also in contrast to the standard case (cf. inset in Fig. 5a).

Fig. 4
figure 4

Temporal evolution of synaptic weights following different stimulation protocols. The plots show the weights within assembly \(A\) for the non-overlapping paradigm. a, b Standard learning/consolidation protocol (order: A-B-C); c, d like the standard protocol but with LTD blocked; e, f extended, interleaved protocol (order: A-B-C-A-C-B-B-A-C-A-B-C-C-B-A). Panels bdf show data from the same simulations as ace but on a longer timescale, revealing consolidation. Data in all panels were averaged over 10 trials; error bands represent the 95% confidence interval, while mostly too small to be visible. Values are given relative to the initial value before learning (\(h_0\)); late-phase weight has been shifted by 1 (cf. Methods) to be visualized accordingly

Fig. 5
figure 5

Scrutinizing the underpinnings of the learning order and overlap effects. a Weight distribution of the synapses within the assemblies after consolidation, revealing that if LTD is blocked, the synapses belonging to the three assemblies are equally potentiated (non-overlapping case). The inset repeats the plot from Fig. 2c for comparison. Weight values were discretized into 100 bins, and the relative frequency was normalized across all bins and all groups of synapses. b Weight distribution of the synapses within the assemblies after consolidation, learned with a modified standard protocol where the breaks between learning A, B, and C last 8 hours instead of 3 seconds (intermediate consolidation). Remarkably, the plot reveals a similar structure as in the no-LTD case. Weight values were discretized into 100 bins, and the relative frequency was normalized across all bins and all groups of synapses. c Overview of the likelihood of avalanches in different organizational paradigms, learned with standard protocol (A-B-C) with LTD blocked. d Overview of the likelihood of avalanches in different organizational paradigms for the intermediate consolidation case (order A-B-C). Data were averaged over 10 trials. Error bars show the 95% confidence interval. Weight values are given relative to the initial value before learning (\(h_0\))

In our model, LTD is caused by medium synaptic calcium levels (cf. Eq. 5), which are evoked by moderate neuronal activity (cf. Eq. 6). Thus, the weakening of pre-existing assemblies seems to originate from their moderate activation evoked by the learning stimulus that acts on the network to form a new assembly. The interleaved protocol employs several times more learning stimuli than the standard protocol to counteract the weakening of individual assemblies (Fig. 4e, f and Suppl. Fig. S5b), which might be the reason why it also enables the attenuation of the learning order effect (cf. Fig. 3d).

Based on these findings, we reasoned that LTD could be the cause of the learning order effect. Therefore, as a next step, we investigated the spontaneous activation in the case of blocked LTD. We found what had already been hinted by the weight distribution (cf. Fig. 5a): the learning order effect indeed vanishes if there is no LTD (Fig. 5c). In addition, the previously observed effect of hub-like structures vanishes. Now, the activation of the hub assembly (B in the “no AC” paradigm, A in the “no BC” paradigm) clearly stands out, as it is expected for an LTP-only model. Furthermore, the overall likelihood of activation increases. In contrast to that, an LTD-only condition would presumably prevent the learning of new memory representations, while stimulation could still cause the weakening of existing memory representations.

Impact of Synaptic Tagging and Capture on Organizational Effects

Next, we investigated the functional consequences of STC with respect to the effects arising from the different organizational paradigms which we characterized in the previous subsection. To this end, we considered the activation of assemblies that had undergone intermediate consolidation after being learned, meaning that the breaks shown in Fig. 1d lasted for 8 hours instead of 3 seconds. Although LTD was switched on for these investigations, much of the learning order effect vanishes, similarly to the no-LTD case (Fig. 5b, d and Suppl. Fig. S5d). The effect of hub-like structures is similar to the no-LTD case as well. It seems that here, LTD is not significantly expressed in the synapses within the assemblies, even though LTD occurs in other synapses (cf. Fig. 5b and inset in Fig. 5a). Apparently, if only one assembly is learned before the next one follows after hours, much of the depression which evokes the learning order effect is too weak to become consolidated. This indicates that late-phase weight change is more than just a low-pass filtered version of early-phase weight change (also cf. Suppl. Fig. S1 and Suppl. Fig. S5). Instead, the criticality of cooperation between synapses to trigger protein synthesis, and of the synaptic tag, renders the transition from early- to late phase weight a nonlinear process.

Comparing the results presented in Fig. 3c and in Fig. 5d shows that different learning/consolidation protocols can lead to qualitatively different results with respect to the effects of learning order and hub-like structure. While our “standard” protocol elicits both substantial late-phase LTP and substantial late-phase LTD, leading to a strong learning order effect and its reduction for assemblies with less overlap, the “intermediate consolidation” protocol rather causes operation in the regime of late-phase LTP only, leading to a weaker learning order effect and promoted activation of hub assemblies. This indicates that the interaction of plasticity processes on different timescales, namely early- and late-phase synaptic plasticity described by STC, enables differential memory effects at the cognitive level.

Priming on Long Timescales

After investigating the functional impact of different organizational paradigms on recency, interference, and enhanced activation through similarity, we provide in this subsection a proof of principle to show that our two-phase plasticity model is also able to describe priming on long timescales. To this end, we use a protocol that includes a brief additional stimulus delivered to all neurons of one selected assembly after learning and consolidation (Fig. 6a). This stimulation causes the respective assembly to be transiently primed for later reactivation, which we investigate by considering the likelihood of avalanche occurrence through background noise as it has been done in the previous subsections. Our results demonstrate that there exists a mechanism for priming enabled by the interplay of early- and late-phase long-term plasticity (Fig. 6b–g). The priming decays on a timescale of minutes to hours, corresponding to the dynamics of early-phase plasticity.

Fig. 6
figure 6

Priming on long timescales enabled by synaptic tagging and capture. a Consolidation and priming phase (following learning as sketched in Fig. 1d but with order A-B-C-A-C-B-B-A-C-A-B-C-C-B-A): consolidation takes place for eight hours. After that, a priming stimulus is applied to one of the assemblies, followed by another period of consolidation and early-phase decay. Then, driven by background noise, spontaneous activation of the memory representations is observed for three minutes. b, c Resulting likelihood of avalanche occurrence in the three assemblies at different times after B has been primed. The values at time zero show the case before (without) priming. d, e Temporal development of the mean total synaptic weight in the assemblies; f, g temporal development of the mean early-phase weight in the assemblies, for the same cases as in b and c. Data in panels b-g are averaged over 10 networks. Error bands show standard error of the mean. Weight values are given relative to the initial value before learning (\(h_0\)). Data for assemblies A and C overlap in panels b, d, f, and g

For investigating priming in the absence of other effects, we sought to attenuate the learning order effect and therefore used the extended interleaved protocol introduced previously. For the non-overlapping case, this approach works fine as it almost entirely eliminates the influence of the learning order (Fig. 6b). However, in the overlapping case (Fig. 6c), the learning order effect still plays a minor role, since in the basal state, A is more active than B and C. Finally, we have to mention that in rare cases, priming stimulation after learning with the interleaved protocol seems to disrupt the recallable memory structure and render reactivation impossible (this occurred in one of ten trials).

The mean total weight shown in Fig. 6d, e exhibits roughly the same qualitative time course as the activation (Fig. 6b, c). Interestingly, in the non-overlapping paradigm, the priming stimulation causes the primed memory representation to behave in an almost attractor-like manner with a likelihood of activation of more than 0.4 (Fig. 6b; also see “Discussion”). It should further be noted that in Fig. 6d, the mean weight in assembly B decreases to a lower value than it had before priming, which may be explained by strong depression of a few synapses in B entering the late phase, while most synapses are subject to early-phase potentiation. Considering the mean early-phase weight (Fig. 6f, g), we find that potentiated weights of assembly B and depressed weights of assemblies A and C correlate with the priming effect of B. Around \(420\,\mathrm {min}\) after priming, the early-phase changes have mostly vanished, while the activation of the assemblies (Fig. 6b, d) has also returned to its baseline level before priming.

Data for more paradigms are shown in Suppl. Fig. S8. The case where only 50% of the assembly neurons receive priming stimulation is shown in Suppl. Fig. S9 (more paradigms in Suppl. Fig. S10). The response to this stimulation leads to the interesting indication that priming can also be mainly driven by LTD. However, this result is not statistically significant, and further investigation on this topic was beyond the scope of this study.

Discussion

In this study, we have shown by proof of principle that synaptic mechanisms of calcium-based plasticity and STC enable to robustly learn, consolidate, organize, and prime long-term memory representations. Our results demonstrate that these features give rise to specific characteristics as they are found in neuropsychological experiments. The robustness of our model against parameter variations is shown by the relative similarity of the results for the standard protocol and the intermediate consolidation protocol (Figs. 3c and 5d), as well as for early- and late-phase weights (Suppl. Fig. S1c and Fig. 3c), and for late-phase weights resulting from randomly initialized weights (Suppl. Fig. S2c).

Measuring the likelihood of spontaneous activation of three cell assemblies in a spiking recurrent neural network after learning and consolidation, we found two factors that guide the activation: 1. the learning order, and 2. the overlap between the assemblies. As an experimental study has shown [66], overlap can account for memory association while not being necessary for the recall of the original separate memories, which matches our findings for the non-overlap and overlap paradigms. There is, moreover, experimental evidence that memories can be potentiated and depotentiated independently of each other [80], which also seems possible in our model. Data from free recall experiments typically exhibit primacy (high probability of activating the first learned items), recency (high probability of activating the last learned items), and contiguity (higher probability of transition between neighboring items) [32, 34]. By our model, we may have provided a mechanistic explanation for the recency effect. However, the direct reproduction of data from free recall experiments that also include the typically weaker primacy and contiguity effects (e.g., [35, 40, 81, 82]) would require a larger number of cell assemblies, which was beyond the scope of this study. Since our detailed network model has a high demand for computational resources, the present study was limited to proof-of-principle investigations with medium-sized networks of 3125 neurons and three assemblies. Nevertheless, in the future, more computational power will enable the simulation of larger networks, which may serve to directly match data from free recall experiments and to provide a mechanistic explanation of primacy and contiguity on a timescale of hours, complementary to previous models [29, 30, 43]. This will, however, require careful tuning of certain plasticity parameters in a biologically plausible way. Related to this, it is important to note that storing more and more memory representations will cause interference and can eventually lead to catastrophic forgetting [83, 84]. It has, however, been shown that multi-phase plasticity can counter catastrophic forgetting [85,86,87]. Studying this with our model could be the goal of a future study serving to explore possibilities of technical application. Another interesting approach to relate to models targeting memory capacity (e.g., [87, 88]) can be to adjust our network such that the memory representations constitute attractors. This would enable the study of transitions between attractor states, which have been proposed to account for the switching between neuropsychological concepts [29, 89, 90], building on the vast amount of previous studies on attractor neural networks. First investigations of attractor dynamics with a slightly modified version of our model have been presented in [91].

In addition to recency in free recall, our model may provide a mechanistic basis for retroactive interference [69,70,71] and enhanced recall of similar memories [72, 73]. Our findings show that it depends on the learning/consolidation protocol whether a hub-like overlap structure enhances or attenuates the activation of particular memory representations (see Figs. 3c and 5d). Furthermore, while overlap causes memory-deteriorating interference in the “standard” protocol (Fig. 3c and Suppl. Fig. S6), it causes memory enhancement in the case of longer intervals between learning stimuli with otherwise equal parameters (“intermediate consolidation” protocol; Fig. 5d and Suppl. Fig. S7). This suggests that in addition to general memory improvement [9], STC can play a functional role in enhancing the retrievability of overlapping memory representations.

From our theoretical investigations, we made the prediction that learning order effects may be essentially caused by LTD (Fig. 5a, c). This can be tested experimentally by a protocol that blocks LTD, for example, via the mGlu receptor antagonist MCPG or via BDNF [92, 93]. Such experiments could, for instance, target the recall of contextual or fear memories (cf. [67, 94]). At the same time, it is important to think about other mechanisms that could disrupt cell assemblies to yield learning order effects. Homeostatic plasticity [95, 96] and synaptic rewiring [97] constitute possible candidates for this, although acting on longer timescales than the plasticity processes that we have considered here.

Our theoretical findings are supported by experiments which have shown that overlap of assemblies acquired on similar timescales as in our “intermediate consolidation” protocol can enhance recall [67, 94]. However, these studies also suggest that memories acquired even further apart in time tend to be separated, which may be investigated in future studies that should consider effects of processes acting on longer timescales. Furthermore, experimental investigation of our “intermediate consolidation” protocol with human subjects in a free recall task may yield a large gain in knowledge. The difference between this protocol and typical free recall protocols is that after learning a memory representation, we wait until it is consolidated (i.e., for about 8 hours) and then learn the next representation, whereas in free recall experiments, the whole list of items is usually learned in short time before consolidation can take place (similar to our “standard” protocol). The experimental investigation of the “intermediate consolidation” protocol would thus enable to test our predictions and yield further interesting insights into the interaction of short- and long-term memory. An issue could be that the subjects would have to stay awake for the whole duration of the protocol, which is about 24 hours, to avoid a biasing effect from sleep consolidation. However, similar experiments have already been done, testing the impact of sleep deprivation on free recall performance [40, 98].

Our study mainly focused on classical Hebbian cell assemblies based on potentiated excitatory synapses. While plasticity of inhibitory synapses, so-called inhibitory plasticity, can enhance excitation–inhibition balance in spiking neural networks [99] and enables functional discrimination between memory representations [100], balanced network states and functional memory recall can be achieved without requiring this type of plasticity (cf. [9, 101]). While it would be very interesting to investigate the influence of inhibitory plasticity on our model, adding such further dynamics will render the simulation and analysis even more complicated.

For our simulations, we assumed early-phase plasticity to be disabled during certain phases. As we pointed out in the “Methods” section, this can be justified by lack of novelty or attention which comes along with lowered neuromodulator concentrations [56, 57], thereby preventing learning (cf. [58, 59]). If, however, early-phase plasticity were enabled all the time, spontaneous activity could exert the following influence on the structure of a cell assembly: either the assembly would stay unaffected, as we have shown for smaller cell assemblies in our previous study [9], or, LTD would cause degradation by downregulating the synaptic weights, or, LTP would cause reinforcement by increasing the synaptic weights. While the latter is not possible due to the firing rate threshold in our model, we assume that spontaneous occurrence of LTD, if any, would not qualitatively change our results. Furthermore, the analysis of our model in [91] shows that input of up to \({\sim }3\,\mathrm {Hz}\) will not elicit early-phase plasticity.

In the final part of our study, we modeled the priming of one of the three previously learned and consolidated memory representations on a timescale of minutes to hours. We demonstrated that this effect can be explained by the interplay between early- and late-phase long-term plasticity as follows: After consolidation, the structure of the memory representations is permanently stored by the late-phase weights, while additional transient information, like priming information, is stored by early-phase weights, which decay across minutes to hours. However, our results also indicate that priming information may “spill over” to the late phase, causing permanent changes in the structure of a memory representation. In our study, the priming effect is strongest if priming is only 10 minutes ago, and it declines until it has vanished almost completely after 7 hours (Fig. 6b, c). This time dependence, as well as the impact of priming in general, may be affected by the strength and duration of the priming stimulus which “loads the representation into short-term memory.” Moreover, the effect of the temporal delay may be influenced by a wide range of further experimental conditions, making it difficult to draw a general conclusion, which has been pointed out by others earlier [102]. Since our results shall constitute a proof of principle, we leave it to further studies to investigate the influence of different kinds of stimulation.

We showed that our model can account for direct positive priming, but our model can likely be extended with relatively low effort to account for other types of priming. We have already encountered predominance of LTD in the case where only 50% of the assembly neurons receive priming stimulation (Suppl. Figs. S9 and S10). While this paradigm still causes a slightly positive priming effect, lowering this fraction even further or simply applying background activity in the presence of early-phase plasticity should revert the outcome and cause a negative priming effect. Such further results would be of great importance, especially because a mechanistic biological theory for negative priming still remains to be found (cf. [46, 103, 104]).

Although various studies have found priming to be relatively independent of the hippocampus [105, 106], other studies have demonstrated a functional relation [107]. Neocortical regions adjacent to the hippocampus have, furthermore, been found to be critical for priming [106]. For this reason, neural networks with hippocampal and neocortical characteristics as the ones presented here are important to understand general processes that could underlie priming. A tuned version of our model might, moreover, account for semantic priming, which is the case if the stimulation of an assembly that overlaps with a second assembly causes primed activation of the second assembly (e.g., in the way that the word “chicken” would prime the word “duck”) [47, 102, 108, 109]. Finally, the fact that the interplay of early- and late-phase plasticity enables priming is another hint that the these two phases might constitute correlates for the concepts of retrieval strength and storage strength [9, 110].

Summary

We have shown that early- and late-phase LTP on the one hand and LTD on the other hand, implemented by calcium-based synaptic plasticity and STC, can critically influence the functionality of memory representations. Our results provide neural correlates to at least partially explain different cognitive effects. In the future, our model could serve as a mechanistic basis for the more detailed reproduction of findings obtained from neuropsychological experiments.