1 Introduction

Living organisms need to make accurate predictions in order to survive (Bubic et al. 2010; Hohwy 2013), posing the question of how do brains learn to make those predictions. Early general models based on classical conditioning (Rescorla and Wagner 1972; Miller et al. 1995), as well as mechanistic models explaining the neural substrate for those predictions (Schultz et al. 1997; Heeger 2017) assume that the prediction errors, changes or future rewards feedback to the predicting neural population, similar to supervised or reinforcement learning paradigms that are common in machine learning. However, recent studies have found that sensory neurons without feedback from higher brain areas encode predictive information (Palmer et al. 2015), a finding that has been supported by simulation studies (Sederberg et al. 2018). This implies that a bottom-up process without explicit feedback—similar to unsupervised learning—should also generate predictions.

In this paper, we present such a mechanism by focusing on postsynaptic latency reduction. This is a well-known effect of spiking time-dependent plasticity (STDP) first mentioned by Song et al. (2000) for a single postsynaptic neuron driven by a specific excitatory input pattern. This effect was explored in detail in a simulation study by Guyonneau et al. (2005) who showed that the latency reduction in the target neuron’s firing time is robust to fluctuations in presyanptic input in the form of jitter and Poissonian background noise. They further analyze the STDP effect on a single neuron receiving fixed (among trials) Poissonian spike trains from each presynaptic neuron and showed that by STDP weights of the earliest afferents will be increased, regardless of the firing rate and level of synchrony of the corresponding neurons. Masquelier et al. (2008) showed how a single postsynaptic neuron under the effect of STDP would learn a single frequent excitatory pattern of spikes even in the presence of a strong background noise and how the firing latency in response to this frequent pattern would decrease over learning trials, a finding that was later extended to rate-modulated Poissonian spike trains (Gilson et al. 2011). In another article Masquelier (2018) quantified the performance of a multi-pattern detector neuron in terms of signal-to-noise ratio and showed that STDP results in an optimal SNR in the response of the neuron when some STDP parameters get tuned, see also (Masquelier and Kheradpisheh 2018). Furthermore, Humble et al. (2012) investigated the effect of STDP on a population of neurons with lateral connections and a global winner-take-all mechanism subjected to a longer spatiotemporal input signal. They showed learning leads to the formation of chains of neurons that are responsive to different patterns in the long time input. Similar results have been reported on learning spatiotemporal patterns by variants of STDP rule in Hunzinger et al. (2012) and Sun et al. (2016).

In this article, we revisit this phenomenon at the micro-level with plastic inhibitory neurons added to the previous setups and analyze the effect of latency reduction at the network level, and finally give it meaning as a computational operation as a mechanism for prediction, adding to the works that show that STDP has important computational roles such as formation of memory traces (Klampfl and Maass 2013) or computing Expectation Maximization for probability distributions (Nessler et al. 2013).

The gist of our argument is that latency reduction implies that neurons fire as early as possible for a given input spike train that is repeated very often; as neurons do not differentiate between a specific stimulus and an early clue of such a stimulus—both being part of seemingly the same input spike train—the neurons can, by STDP, fire earlier than the stimulus itself. Furthermore, we expand on the previous studies focused on excitatory neurons to include inhibition and illustrate the parameter regime in which inhibitory plasticity is compatible with latency reduction. However, the latency reduction mechanism has other uses in terms of neural code. First, as neurons fire as early as possible when a stimulus is presented, their spikes will concentrate in a small time window, and thus they are easier to decode. Second, we show that the latency reduction can also lead to a reduction in the number of spikes, which translates as a reduction in metabolic costs for encoding frequent stimuli.

We develop our argument by studying simple models of neurons subject to fixed input spike trains. We use a combination of simulations and mathematical analysis to derive our results, starting from the evolution of a single postsynaptic spike at very short timescales we expand to larger scales that conclude in the emergence of predictions and efficient code at the level of populations of neurons in large timescales.

The rest of this paper is organized as follows. First, we present the models of neurons and STDP in Sect. 2. Second, we study the effects of STDP in a single postsynaptic spike in very small timescales \(\sim 10 ms\), focusing on latency reduction and the reduction of the number of postsynaptic spikes in Sect. 3. In Sect. 4, we expand those results to long postsynaptic spike trains , finding that STDP forces postsynaptic neurons to fire only once at the onset of the presynaptic spike train. Afterward, we provide an interpretation of this spike concentration in terms of neural code performance, showing that it leads to lower number of spikes and synchronization. We finalize by illustrating that the same mechanism of latency reduction leads to encoding predictions in Sect. 5.

2 Models

2.1 Leaky integrate-and-fire neuron

Neurons are considered to be the basic computational units in the nervous system. Their main feature is the capacity to receive information through electrical impulses, combine this information and send impulses to other neurons. In this paper, we model them as leaky integrate-and-fire neurons with a refractory period (Lapique 1907). In this model, the state of a neuron at a given time is described by its membrane potential v(t), which evolves according to the equation

$$\begin{aligned} \tau _m \dfrac{\mathrm{d} v(t)}{\mathrm{d}t} = -(v(t) - v_0) + i(t), \end{aligned}$$
(1)

where \(\tau _m = 10\,\)ms, \(v_0 = -70\) mV. i(t) is the input to the neuron at time t. When the membrane potential reaches a certain threshold \(v_{th} = -50\,\)mV, the neuron “fires” or “spikes,” meaning that it emits a pulse of current. After firing, the membrane potential is reset to its resting state \(v_0\) and kept frozen at this value for a fixed period of time called the refractory period \(t_{\mathrm{ref}} = 1\,\)ms.

The firing of a neuron generates pulses of current that arrive at other neurons, which in turn update their membrane potentials. If neuron a receives the spikes of neuron b we will say that there is a synapse going from the second to the first. The receiving neuron is called postsynaptic and the sending neuron is the presynaptic one. This synapse is characterized by a weight \(w_{ab}\) and a delay \(d_{ab}\) which correspond, respectively, to the gain and the latency that the pulse of neuron a goes through before arriving at b.

2.2 Input spike trains

Neurons communicate mainly through action potentials or spikes, which are typically modeled as Dirac delta functions, hence the input to a neuron can be described as

$$\begin{aligned} i(t) = \sum _n w_n\delta (t-t_n), \end{aligned}$$
(2)

where \(w_n\) is the weight of the spike, which corresponds to the strength of the synapse from which the spike comes, and \(t_n\) is the arrival time of the spike. The weights of the synapses can be positive, if the presynaptic neuron is excitatory, or negative, if it is inhibitory. Through this paper we will assume that every neuron gets an input that will be repeated, meaning that a neuron will always get spikes from different synapses, and although the weights of the synapses might change, the times \(t_n\) of the spikes will remain the same in every input repetition. Each synapse comes from a presynaptic excitatory neuron with probability 0.8 or an inhibitory one with probability 0.2.

2.3 Spiking time-dependent plasticity

Networks of neurons learn by modifying the strength of the connections between them. There is a rich literature on what rules those weights follow in biological neurons and their respective implications (Dayan and Abbott 2001). For the purposes of this paper, the neurons will adapt their connections according to the spiking time-dependent plasticity (STDP) paradigm (Sjöström and Gerstner 2010; Gerstner et al. 1996).

In STDP, the weight of a connection is modified depending on the time interval between pairs of pre- and postsynaptic spikes. For every pair, the weight of the synapse is changing according to the equations

$$\begin{aligned} \Delta w (\Delta t) = {\left\{ \begin{array}{ll} A_{+}(w) \mathrm{e}^{-\frac{|\Delta t|}{\tau _{s}}} \quad \text { if }\Delta t \ge 0\\ -A_{-}(w) \mathrm{e}^{-\frac{|\Delta t|}{\tau _{s}}} \quad \text { if } \Delta t < 0 \end{array}\right. } \end{aligned}$$
(3)

where \(\Delta t = t_{\mathrm{post}} - t_{\mathrm{pre}}\) is the time difference between the postsynaptic spike and the presynaptic one, \(\tau _{s} = 20\,\)ms. Based on previous works (Werner and van Hemmen 2000; Van Rossum et al. 2000), we define \(A_+\) and \(A_-\) as

$$\begin{aligned} \begin{aligned} A_+(w)&= \eta _+ (w_{\max }-w),\\ A_-(w)&= \eta _- (w-w_{\min }) \end{aligned} \end{aligned}$$
(4)

where \(\eta _- = 0.015,\ \eta _+ = 0.01\), \(w_{\max }^e = 10\,\)mV and \(w_{\min } = 0\). Inhibitory synapses follow the same rules as their excitatory counterparts but with parameters \(\eta _- = 0.045\) , \( \eta _+ = 0.03 \) and \(w_{\min }^i = -20\,\)mV. This inhibitory kernel has been experimentally observed (Vogels et al. 2013) and its symmetry with respect to its excitatory counterpart will make our analysis simpler, although we shall explore other kernels in “Appendix B”.

2.4 Model limitations and required features

We must note that the models used here are heavy simplifications of real neurons. LIF neurons do not exhibit the rich range of dynamics that real neurons possess(Izhikevich 2004), ion channel kinetics are more complicated than simple Dirac deltas (Chapeau-Blondeau and Chambet 1995) and the STDP model used here cannot account for the evolution of synaptic weights when the frequency of postsynaptic or presynaptic spikes are high (Pfister and Gerstner 2006). However, those models contain the main biologically realistic features that we need for the rest of this study. First, the time constants of the neuron membrane potentials(Gerstner et al. 2014) and the STDP interactions (Bi and Poo 1998) are at least an order of magnitude smaller than the duration of the input spike trains associated to biologically realistic stimuli-evoked spatiotemporal patterns (Rolston et al. 2007; Prut et al. 1998). Second, the neurons have a low firing rate (Roxin et al. 2011). Third, the synapses whose spikes presynaptic spikes arrive shortly before a postsynaptic spike get reinforced, while those arriving afterward get depressed (Sjöström et al. 2008; Pfister and Gerstner 2006). Finally, the homeostatic consideration that firing rates of neurons should not increase widely, which is a natural requirement on metabolic grounds (Turrigiano and Nelson 2004) can easily be incorporated by the depressive term \(A_-\). Thus, we will keep these well-known models (Gerstner et al. 2014) on the grounds that they are analytically tractable and qualitatively plausible.

3 Evolution of a single postsynaptic spike

In this section, we show that STDP can change individual postsynaptic spikes by reducing their latencies and their number. We will start by presenting simple scenarios with excitatory inputs in which both effects are easy to illustrate, then show how inhibitory synapses can be added to the model, and finally show that those effects can appear in random input spike trains by presenting simulations. It is worth noticing that the time windows in this section are on the order of \(\tau _s\) and the number of repetitions of each input pattern will be small.

3.1 Latency reduction

If a fixed train of presynaptic spikes is repeated very often, then the spikes that arrive before the postsynaptic spike get reinforced. This implies that the postsynaptic spike might then be triggered earlier (Song et al. 2000; Gerstner et al. 1996). When this happens, the refractory period of the postsynaptic neuron would prevent a second spike on the original spiking site. However, when the postsynaptic spike happens earlier and earlier, it might lead to a proliferation of spikes by having a new spike appear at the time of the original postsynaptic spike. Following previous literature (Song et al. 2000; Abbott and Nelson 2000; Kempter et al. 2001), to prevent this effect, we assume that long-term depression—the weakening of synaptic weights—is stronger than long-term potentiation—the strengthening of postsynaptic weights.

This is easy to understand in a simple scenario: Considering a very long, excitatory presynaptic spike train which generates a single postsynaptic spike at some time \(t_0\). The postsynaptic spike will advance through the spike train, and after some repetitions it will be triggered one presynaptic spike earlier. After this advancement is repeated many times, the postsynaptic spike is triggered at time \(t_\infty \), very far (in time) from the place where it was first triggered, so that

$$\begin{aligned} t_\infty \ll t_0. \end{aligned}$$
(5)

The membrane potential decays exponentially, meaning that the effect of the postsynaptic spike at time \(t_\infty \) on the membrane potential is of order \(O(\mathrm{e}^{-\frac{t_0 - t_\infty }{\tau _m}})\), which is negligible. Thus, the membrane potential at time \(t_0\) is now only dependent on the presynaptic spikes that are close. If those presynaptic spikes have been left as they where by the passage of the postsynaptic spike, then a new postsynaptic spike will be generated at time \(t_0\). To avoid the appearance of this postsynaptic spike it is therefore necessary that the passage of the postsynaptic spike weakens the presynaptic ones. We illustrate this point in Fig. 1 with the functions and parameters that we will use in subsequent sessions.

Fig. 1
figure 1

Latency reduction and spike proliferation: We plot the membrane potential (left) and firing times (right) of a postsynaptic neuron that receives a constant train of spikes with inter-spike interval of \(3.5\,\)ms and strength \(5.5\,\)mV, from time \(t=0\,\)ms to \(t=150\,\)ms, and we add an extra spike at \(t=150\,\)ms with potential \(2\,\)mV. The neuron generates a single postsynaptic spike at the original input presentation (Repetition 0). The upper plots reflect the case \(\eta _+=\eta _-\), while for the lower ones we picked \(\frac{3}{2}\eta _+ = \eta _-\). After an initialization period, the postsynaptic spike moves forward in time at a constant rate. As this happens, a single presynaptic spike will get reinforced proportionally to the \(\eta _+\) and dampened proportionally to \(\eta _-\). If LTP is equal to LTD, after the postsynaptic spike happens much earlier than before, the membrane potential of the postsynaptic neuron will reach the threshold again. This second postsynaptic spike would move forward in time at the same speed as the strengths of the spikes are left unchanged by the compensation of LTD and LTP (upper plots). In the case where \(\eta _+<\eta _-\), the depression compensates the potentiation, so there is no second postsynaptic spike

Note that the argument that we give here is qualitative in nature, in the sense that we simply state that LTD should dominate LTP through the constant \(\eta \), but we have not studied how to find that ratio. As this would depend on the exact parameters of the regular spike train—and thus would not be directly generalizable—we will simply assume that the brain operates in a parameter regime in which spikes do not proliferate.

3.2 Late spike disappearance through synaptic noise

If latencies might be reduced, then two postsynaptic spikes that are triggered at distant points in time might become closer in subsequent learning trials. We must then ask what happens to a pair of postsynaptic spikes that occur very close in time. In this section, we show that in the absence of synaptic noise the two spikes can coexist, but random modifications of the presynaptic weights—induced, for instance, by other presynaptic inputs—can lead to the disappearance of the second postsynaptic spike.

There are many possible scenarios that we might consider when we have pairs of postsynaptic spikes in the same neuron: We must consider the time between the two spikes, the movements in time of both of them and the possibility of synaptic noise. The case when two postsynaptic spikes happen originally very close in time is extremely rare—because postsynaptic spikes are sparse. The case where the first postsynaptic spike also moves is not interesting, because the spike will move forward in time, increasing the distance between the two postsynaptic spikes and thus reducing the LTD effect on the second spike—note that the second spike would not move as fast because the presynaptic spikes between them would be depressed by the first. Therefore, we will consider the case where there is an early postsynaptic spike at some fixed time that will remain in place, and a second postsynaptic spike that will initially be triggered very far in time.

The intuition here is that there is a time interval for the second postsynaptic spike, in which the LTD of the first postsynaptic spike would lead to a decrease in the membrane potential of the postsynaptic neuron at the time of the second postsynaptic spike, which could lead to the irreversible disappearance of the second postsynaptic spike or its recession. Outside of this time interval, the second postsynaptic spike will reduce its latency, approaching the early postsynaptic spike and the disappearance zone. In the remaining of this section, we will show that this interval is never reached in a deterministic system but that the addition of noise can enforce this disappearance.

Consider a long presynaptic spike train with presynaptic spikes arriving at \(t_0, t_1,\ldots t_N\), which generates two postsynaptic spikes, one at time \(t_0\), which is fixed and will appear at every presentation of the spike train, and another one that is originally triggered at \(t_N\). For the second spike to disappear, it can either do so at \(t_N\) or first advance through the spike train—that means, being triggered at \(t_{N-1}\), then at \(t_{N-2}\) and so on—and eventually die. For now, we assume that \(t_N - t_0 \gg \tau _s\), so that initially the spike at time \(t_N\) evolves independently of the spike at time \(t_0\), and it would not disappear at \(t_N\). Consider now that the input has been repeated long enough so that the second postsynaptic spike is now triggered at \(t_i\), and the effects of the STDP generated by the spike at \(t_0\) are not negligible to the presynaptic weight \(t_{i-1}\), which is associated to the presynaptic spike at \(t_{i-1}\). If the postsynaptic spike is originally triggered at \(t_i\), then it would move to \(t_{i-1}\) only if, after repeating the same input many times,

$$\begin{aligned} v(t_{i-1}) = \sum _{k=1}^{i-1} w_k \mathrm{e}^{-\frac{t_k-t_{i-1}}{\tau _m}} \ge v_{th}. \end{aligned}$$
(6)

After \(v(t_{i-1})\) crosses the \(v_{th}\) threshold, the postsynaptic spike at \(t_i\) moves to \(t_{i-1}\), and thus the time difference between every presynaptic spike at \(t\le t_{i-1}\) and the postsynaptic spike is reduced. This naturally implies that the synaptic weights \(w_k\) for all \(k\le i-1\) increase, thus the postsynaptic spike cannot disappear because the membrane potential at \(v(t_{i-1})\) cannot decrease unless the postsynaptic spike moves to \(t_{i-2}\). Therefore, with a deterministic input the second postsynaptic spike precedes and stops at an equilibrium distance from the first one. This implies that the weights of the presynaptic neurons arriving at each specific time point in the interval would also reach their equilibrium values.

This argument assumes that presynaptic spike trains are always repeated with fixed spike timings but with weights that are affected by LTP and LTD. This is generally not true, as there are many factors that can introduce stochasticity on the evolution of the weights, such as jitter, the stochastic nature of molecular dynamics on the synaptic cleft and on the neuron membrane.

If we now consider the stability of both postsynaptic spikes with respect to that noise, we easily realize that they are not equal: While the presynaptic spikes that generate the first postsynaptic spike are only subject to LTP and noise, the presynaptic spikes that generate the second spike—which happen necessarily between postsynaptic spikes—are subject to both LTP—from the late postsynaptic spike—and LTD—from the earlier postsynaptic spike—on top of the noise.

This difference implies that the noise can make a postsynaptic spike disappear or recede, either by directly weakening the associated presynaptic weights or strengthening them, so that the postsynaptic spike moves into a region where LTD dominates and it would be later erased or pushed back.

To explain this in the setting that we used before, consider a neuron with a postsynaptic spike at time \(t_i\) that would not move to \(t_{i-1}\) in the previous deterministic system. However, now the weights evolve by the combined effects of that spike, an earlier postsynaptic spike at time \(t_0\) and some noise. The membrane potential at time \(t_i\) and after r repetitions of the input spike train follows

$$\begin{aligned} v(t_{i}) = \sum _{k=1}^{i} w_k \mathrm{e}^{-\frac{t_k-t_{i}}{\tau _m}} + \xi _{t_{i}}, \end{aligned}$$
(7)

where \(\xi _t\) is the contribution of the random evolution of the weights to v(t) given by

$$\begin{aligned} \xi _{t_i} = \sum _{k=1}^i \delta w_k \mathrm{e}^{-\frac{t_k-t_{i}}{\tau _m}} \end{aligned}$$
(8)

where \(\delta w_k\) is the deviation of weight \(w_k\) from its deterministic evolution; in the case of Gaussian noise, for instance, it would lead to an Ornstein–Uhlenbeck process for the evolution of \(v(t_i,r)\) across the number of trial repetitions r. Note that the noise is a variable reinitialized at every repetition, but its effects on the weights remain across repetitions.

If this postsynaptic spike train is repeated very often, the deterministic part of the weights goes to a fixed value, which is small for \(k>i\) and thus \(v(t_k)\sim \xi _{t_k}\) for all \(k>i\). Thus, under the assumption

$$\begin{aligned} \xi _{t_{i}} < v_{th}- \sum _{k=1}^{i} w_k \mathrm{e}^{-\frac{t_k-t_{i}}{\tau _m}} \end{aligned}$$
(9)

in a specific trial the second spike will be absent. Subjected to the ever present postsynaptic spike at \(t_0\), the weights \(w_k\) will decrease for all values of k after this trial makes the neuron less likely to fire in the subsequent trials. This negative drift will finally lead to the irreversible disappearance of the postsynaptic spike at \(t_i\) or its delay. This is illustrated in Fig. 2.

Fig. 2
figure 2

Noise deletes a late spike in a regular presynaptic spike train: We plot the membrane potential (left) and firing times (right) of a postsynaptic neuron that receives a constant train of spikes with inter-spike interval of 5 ms and strength \(7.5\,\)mV, from time \(t=0\,\)ms to \(t=150\,\)ms. We add an extra spike at \(t=0\,\)ms with potential 5mV, which forces a postsynaptic spike at time 0.5 ms. The top row is deterministic, while the bottom row is stochastic as the weights are subject to Gaussian noise with variance of 0.3. Note that, during its existence, the latency of the postsynaptic spike subject to noise decreases faster than its noiseless counterpart

3.3 Generalization to inhibitory plasticity

Until now we have only considered excitatory neurons. However, in biological systems, inhibitory synapses are also present and show plasticity (Vogels et al. 2013). Naturally, this might compromise the effects described in the previous section, as an inhibitory synapse that gets potentiated could counteract the effects of excitatory STDP. For instance, it might decrease the membrane potential and thus increase the latency of the postsynaptic neuron (Effenberger et al. 2015). Our goal in this section is to find the parameter regime in which the presence of inhibitory plasticity does not compromise the latency decrease and, by extension, the disappearance of postsynaptic spikes.

Intuitively, as long as the STDP in inhibitory synapses is weaker than the STDP in excitatory ones, the latency of postsynaptic spikes would still decrease. The question is then to find a way of measuring “how much weaker” it has to be. To address this issue, we must find a boundary parameter set for inhibitory synapses that guarantees that latency would be reduced, and then we can simply take any parameter set that is between this boundary parameter set and the only excitatory STDP.

To identify the parameter regime in which latency reduction for a single spike appears, we assume that the STDP keeps the balance between excitation and inhibition, in the sense that the average input to a single neuron is maintained constant (Brunel 2000). To maintain this balance, the potentiation of excitatory synapses is compensated by the potentiation of inhibitory synapses. Potentiating all synapses but maintaining the average input leads to the increase in fluctuations of the membrane potential, meaning that the membrane potential preceding a postsynaptic spike would change more around the average, and thus it can still lead to an earlier postsynaptic spike.

Consider a single postsynaptic spike at time \(t_{\mathrm{post}}\). For \(t<t_{\mathrm{post}}\),

$$\begin{aligned} v(t) = \sum _{t_k<t} w_k \mathrm{e}^{-\frac{t-t_k}{\tau _m}}, \end{aligned}$$
(10)

and initially \(v(t) < v_{th}\). Now we wonder what happens when the weights \(w_k\) change, in particular whether the postsynaptic spike will advance, recede or disappear. This depends on the exact values of \(w_k\) and \(t_k\), so to make more generic statements we are interested in the value

$$\begin{aligned} {\mathbb {E}}_r\left[ \Delta t_{\mathrm{post}}\right]= & {} {\mathbb {E}}\left[ t^r_{\mathrm{post}} - t_{\mathrm{post}}\right] \nonumber \\= & {} \left( {\mathbb {E}}\left[ t^r_{\mathrm{post}}\right] -t_{\mathrm{post}}\right) \text {Pr}\left[ \exists s\right] \end{aligned}$$
(11)

where r accounts for the number of times that the spike train has been repeated, and \(\text {Pr}\left[ \exists s\right] \) is the probability that a postsynaptic spike still exists, and the expectations are taken over the presynaptic spike trains—a list of tuples \((w_k,t_k)\) sampled from some predefined distribution—that generate a postsynaptic spike at time \(t_{\mathrm{post}}\). In simpler words, we are trying to calculate whether the postsynaptic spike is expected to move forward (\({\mathbb {E}}\left[ \Delta t_{\mathrm{post}}\right] (r)< 0\)) or backward (\({\mathbb {E}}\left[ \Delta t_{\mathrm{post}}\right] (r)> 0\)), ignoring the ones that disappeared, if we only have some information about the distribution from which the list of \((w_k,t_k)\) was sampled.

We know that increasing the input excitatory weights can only lead to an earlier postsynaptic spike, because v(t) can only increase and thus it might reach \(v_{th}\) earlier. We will take this a step further and assume that this statement is also true about the average weights, meaning that when the expected input increases, the expected postsynaptic firing time decreases. In more formal terms, we are assuming that \({\mathbb {E}}_r\left[ \Delta t_{\mathrm{post}}\right] \) is a function that decreases monotonically with

$$\begin{aligned} {\mathbb {E}}\left[ \Delta _r v(t)\right]= & {} {\mathbb {E}}\left[ \int _{-\infty }^{t}\Delta _r i(t)\mathrm{e}^{-\frac{t-x}{\tau _m}}\mathrm{d}x\right] \nonumber \\= & {} \int _{-\infty }^{t}{\mathbb {E}}\left[ \Delta _r i(t)\right] \mathrm{e}^{-\frac{t-x}{\tau _m}} \mathrm{d}x, \end{aligned}$$
(12)

for all \(t<t_{\mathrm{post}}\), meaning that if the expected value of \(\Delta v(t)\) averaged over all realizations of the input spike train producing a spike at \(t_{\mathrm{post}}\) is positive, then \({\mathbb {E}}_r\left[ \Delta t_{\mathrm{post}}\right] \) will be negative.

This assumption, albeit natural, requires some careful consideration. Specifically, we must clarify the distribution over which the expectations are taken, which corresponds to all possible presynaptic spike trains shortly preceding a postsynaptic spike. Those spike trains have fixed timings for every postsynaptic spike under consideration, but are updated systematically because postsynaptic spikes evolve with the input repetitions and the noise. Thus, this distribution considers samples in which a new spike has just appeared or samples where a postsynaptic spike has recently been displaced by a short time.

The subsequent step is to find the conditions that guarantee that \({\mathbb {E}}\left[ \Delta _r v(t)\right] \) increases. A sufficient condition for this to happen is to have

$$\begin{aligned} {\mathbb {E}}\left[ \Delta _r i(t)\right] = \Delta _r {\mathbb {E}}\left[ i_e(t)\right] - \Delta _r{\mathbb {E}}\left[ i_i(t)\right] > 0,\ \forall t< t_{\mathrm{post}}\nonumber \\ \end{aligned}$$
(13)

where \({\mathbb {E}}\left[ i_e(t)\right] \) is the expected input to the neuron at time t, and \( {\mathbb {E}}\left[ i_i(t)\right] \), \({\mathbb {E}}\left[ i_i(t)\right] \) is simply its decomposition in inhibitory and excitatory inputs, which gives us

$$\begin{aligned} \begin{aligned} {\mathbb {E}}\left[ i_e(t)\right] = \rho _e\int _{0}^{\infty } \mu _{w_e}(w,t) \mathrm{d}w\\ {\mathbb {E}}\left[ i_i(t)\right] = \rho _i\int _{0}^{\infty } \mu _{w_i}(w,t) \mathrm{d}w \end{aligned} \end{aligned}$$
(14)

where \(\rho _e, \rho _i\) are the rates of incoming spikes and \(\mu _{w_e}(w,t), \mu _{w_i}(w,t)\) the probabilities of the weights associated to time t.

Thus, to maintain the condition from Eq. 13, we must ensure that the parameters \(\mu _{w_e}\), \(\mu _{w_i}\), \(\eta _+^e\), \(\eta _+^i\), \(w_{\min }^e\), \(w_{\min }^i\) are such that

$$\begin{aligned}&\rho _e\int _{0}^{\infty } \Delta w_e(r) \mu _{w_e}(w,t) \mathrm{d}w\nonumber \\&\quad > \rho _i\int _{0}^{\infty } \Delta w_i(r)\mu _{w_i}(w,t) \mathrm{d}w, \end{aligned}$$
(15)

where \(\Delta w(r)\) are given by the STDP Eq. 3 over many repetitions—counted by r—of the input spike train. We will now find a parameter regime in which this holds by finding its boundary. In other words, we are interested in the parameter set in which the excitatory increase in weight exactly matches the inhibitory increase in weight, which for the time constants of inhibition being equal to that of excitation leads us to the condition

$$\begin{aligned}&\rho _e\int _{0}^{\infty } \Delta w_e(r) \mu _{w_e}(w,t) \mathrm{d}w\nonumber \\&\quad = \rho _i\int _{0}^{\infty } \Delta w_i(r)\mu _{w_i}(w,t) \mathrm{d}w. \end{aligned}$$
(16)

Note that it is not enough to find two weight distributions \(\mu _{w_e}\), \(\mu _{w_i}\) where

$$\begin{aligned}&\rho _e\int _{0}^{\infty } A_+^e(w_e) \mu _{w_e}(w,t) \mathrm{d}w \nonumber \\&\quad = \rho _i\int _{0}^{\infty } A_+^i(w_i)\mu _{w_i}(w,t) \mathrm{d}w, \end{aligned}$$
(17)

because this would only work for the first input repetition. We have to ensure that even after STDP changes the distribution, the equality holds. Since there are typically fewer inhibitory synapses than excitatory ones, we correct the input rates and STDP parameters by the ratio

$$\begin{aligned} \alpha = \dfrac{\rho _i}{\rho _e} \end{aligned}$$
(18)

that is also intrinsic to the probability distributions

$$\begin{aligned} \alpha \mu _{w_i}\left( \alpha x,t\right) = \mu _{w_e}(x,t) \ \forall x,t, \end{aligned}$$
(19)

and the STDP parameters

$$\begin{aligned} \alpha A_+^i(\alpha x) = A_+^e(x)\ \forall x. \end{aligned}$$
(20)

If these properties are satisfied,by a simple change of variable, we can show that

$$\begin{aligned} \begin{aligned}&\rho _e\int _{0}^{\infty } A_+^e(x) \mu _{w_e}(x,t) \mathrm{d}x \\&\quad = \dfrac{1}{\alpha } \rho _i \int _{0}^{\infty } \alpha A_+^i(\alpha x) \alpha \mu _{w_i}\left( \alpha x,t\right) \frac{1}{\alpha } \mathrm{d}(\alpha x)\\&\quad = \rho _i\int _{0}^{\infty } A_+^i(y)\mu _{w_i}(y,t) \mathrm{d}y. \end{aligned} \end{aligned}$$
(21)

Furthermore, if we take a pair of inhibitory and excitatory weights such that \(w_e = \alpha w_i\) we have that after applying the STDP rule,

$$\begin{aligned} \alpha w_i \rightarrow \alpha (w_i + A_+^i(w_i))= & {} \alpha w_i + \alpha A_+^i(\alpha w_e) \nonumber \\= & {} w_e + A_+^e( w_e) \leftarrow w_e, \end{aligned}$$
(22)

meaning that the weight probability changes in such a way that

$$\begin{aligned} \mu _{w_e}'\left( x + A_+^e(x),t\right)= & {} \mu _{w_e}\left( x,t\right) = \alpha \mu _{w_i}\left( \alpha x,t\right) \nonumber \\= & {} \alpha \mu _{w_i}'\left( \alpha \left( x + A_+^i(x)\right) ,t\right) , \end{aligned}$$
(23)

where \(\mu _{w_e}'\) and \(\mu _{w_i}'\) are the weight distributions after STDP has acted once. Thus, if Eq. 19 holds at some point, it will also hold for all subsequent iterations of the input spike pattern.

Thus, we have found a set of conditions that satisfy Eq. 19 at \(r=0\) and for any subsequent \(r>0\) for the case where the postsynaptic spike does not change during the r repetitions. Notice that the self-consistency of this condition does not make any assumptions about the learning constant or the \(\Delta t\) dependent term on STDP, or even its sign, it only requires that the expected increase—or decrease—in excitatory input is matched by the expected increase—or decrease—in inhibitory input. In particular, this symmetry does not change if the postsynaptic spike advances, because the STDP kernel has the same ratio of potentiating inhibitory and excitatory synapses. In other words, when a postsynaptic spike changes places before the rth input repetitions, the variance of the input before the postsynaptic spike still increases and, conversely, the variance after the postsynaptic spike decreases.

Now, we have a large set of parameters in which latency reduction is expected to happen. Any STDP parameters for which \(\alpha A_+^i(\alpha x) < A_+^e( x)\) combined with Eq. 19, or distribution of weights with \(\alpha \mu _{w_i}\left( \alpha x,t\right) < \mu _{w_e}(x,t)\) with Eq. 20, or both cases combined.

It is worth noticing that in the case when all the equalities Eqs. 19 and 20 are met, we would still expect the latency to decrease. The reason is that even if

$$\begin{aligned} {\mathbb {E}}\left[ \Delta _r v(t)\right] = {\mathbb {E}}\left[ \Delta _r i(t)\right] = 0, \end{aligned}$$
(24)

the variance of v(t) increases. More explicitly,

$$\begin{aligned} \begin{aligned} \Delta _r\text {Var}\left[ v(t)\right]&= \Delta _r \int _{-\infty }^{t}\text {Var}\left[ i(t)\right] \mathrm{d}t\\&= \Delta _r \int _{-\infty }^{t}\left( {\mathbb {E}}\left[ i^2(t)\right] - {\mathbb {E}}\left[ i(t)\right] ^2\right) \mathrm{d}t \\&= \Delta _r \int _{-\infty }^{t}{\mathbb {E}}\left[ i^2(t)\right] \mathrm{d}t = \int _{-\infty }^{t} \Delta _r {\mathbb {E}}\left[ i_\mathrm{e}^2(t)\right] \mathrm{d}t \\&\quad + \int _{-\infty }^{t} \Delta _r {\mathbb {E}}\left[ i_i^2(t)\right] \mathrm{d}t \end{aligned}\nonumber \\ \end{aligned}$$
(25)

where the term \({\mathbb {E}}\left[ i(t)\right] ^2 = 0\) by the symmetry of the weights and it is maintained at zero by the symmetry of the STDP. Since we are only concerned with \(t<t_{\mathrm{post}}\), STDP potentiates both inhibitory and excitatory synapses, so

$$\begin{aligned} \begin{aligned} \Delta _r {\mathbb {E}}\left[ i_i^2(t)\right] ,\Delta _r {\mathbb {E}}\left[ i_\mathrm{e}^2(t)\right] > 0 \end{aligned} \end{aligned}$$
(26)

and therefore the variance increases. Naturally, if the variance of a certain distribution increases while keeping its mean constant, then the probability of reaching a value higher than some threshold—\(v_{\mathrm{th}}\)—also increases.

The approach outlined here can be also used for other STDP kernels. While the symmetry in the excitatory and inhibitory STDP kernels might not exist for some choices of inhibitory and excitatory plasticity, the approach can, in principle, still be used by making sure that the mean or the variance of the inputs to a neuron would grow before each postsynaptic spike

The nonproliferation of spikes can be derived by a similar argument, although in this case the mean or the variance (or both) of the presynaptic input to the postsynaptic neuron will decrease due to the depressive nature of STDP for \(t>t_{\mathrm{post}}\). In general, the idea of having the depression stronger than the potentiation would still work, as long as the depression of inhibitory synapses is weaker or equal than that of excitatory synapses. As this calculation is essentially the same as the one we just presented, we will skip it.

3.4 Numerical verification for random input spike trains

The examples presented to illustrate the latency reduction and the disappearance—or delay—of late postsynaptic spikes were simple, so we must now extend them to a more general case. To do so, we simulated spike trains where the times of the presynaptic spikes are randomly sampled at the beginning and then fixed for every subsequent repetition, including only excitatory or excitatory and inhibitory STDP, noise and the presence of an earlier postsynaptic spike. The results are presented in Table 1 and agree with our previous conclusions: A single postsynaptic spike tends to reduce its latency, if there are multiple postsynaptic spikes in a short time window the later ones tends to disappear, and the presence of noise increases those effects. Note that we have not included jitter or probabilistic presynaptic spikes, choosing instead to have noise directly on the weight evolutions. As both cases have been addressed before (Guyonneau et al. 2005) with similar conclusions, we shall not repeat them here.

Table 1 Effects of STDP on short random spike trains: We explored the effects of STDP on the postsynaptic spike train of a neuron receiving 8 excitatory and 2 inhibitory presynaptic spikes arriving at uniformly sampled times on the interval \(\left[ 0, 40\,\mathrm{ms}\right] \) and the stimulus is repeated 100 times

So far we have only considered effects on small time scales, meaning that there were only a few spikes on a time interval of the order of \(10\,\)ms, and the postsynaptic spike train would evolve over a few repetitions, on the order of \(20\,\)ms. This leads us to the conclusion that, with plausible assumptions on the parameters of our model, an individual postsynaptic neuron will fire a specific postsynaptic spike earlier after many repetitions of the same presynaptic spike train and that if two postsynaptic spikes are close in time, then the later one could disappear.

4 Postsynaptic spike train

Now, we study the effects of the previously described phenomena, which act on small temporal scales and affect only one or two postsynaptic spikes, for a population of postsynaptic neurons, each one receiving many presynaptic spike trains happening over time scales much larger than \(\tau _m\) or \(\tau _s\). Specifically, we will explore the latency reduction and suppression or delaying of late postsynaptic spikes and the change in the postsynaptic spike distribution.

Before studying those effects, we must validate some of the assumptions that we made in the previous section. In particular, we assume that all the input spikes came from different synapses, which allowed us to treat the weights of all presynaptic spikes as independent. This is a valid assumption when we are considering short time intervals, as the sparsity of presynaptic firing and the existence of refractory periods implies that a single synapse would typically not fire more than once during a short presynaptic spike train. However, when there is a long presynaptic spike train, a presynaptic neuron might contribute to that spike train more than once, thus our assumption might be invalid and the phenomena described in the previous section might not appear. To ensure that the phenomena of latency reduction and late spike disappearance are still present in long spike trains, we use a combinatorial argument and count the number of synapses that might evolve in a non-trivial fashion, which we present in “Appendix A”.

We can now consider the first time that an input presynaptic spike train is presented. Every neuron starts at \(v(0) = 0\) and then its membrane potential will change depending on its inputs. As the input spike train consists of independent spikes with independent weights, the times of the first spike have a probability distribution \(f_0^{1}(t)\) with support on \(t>0\), which depends on the parameters of the input spike train. After spiking, every neuron resets its membrane potential to zero, and thus the distribution of inter-spike intervals \(f_0^{\mathrm{ISI}}(t)\) follows

$$\begin{aligned} f_0^{\mathrm{ISI}}(t) = f_0^{1}(t-t_{\mathrm{ref}}). \end{aligned}$$
(27)

After the input has been repeated many times, the distribution of postsynaptic spikes changes to \(f_\infty ^1\) and \(f_\infty ^{\mathrm{ISI}}\), respectively. Specifically, the first spikes reduce their latency on average and thus move closer to \(t=0\), while the inter-spike intervals increase, due to the depressive effect of postsynaptic spikes that repels or eliminates late postsynaptic spikes. Therefore,

$$\begin{aligned} \begin{aligned} F_\infty ^{1}(t)&= \int _{0}^{t} f_\infty ^{1}(x) \mathrm{d}x \ge \int _{0}^{t} f_0^{1}(x)\mathrm{d}x = F_0^{1}(t)\\ F_\infty ^{\mathrm{ISI}}(t)&= \int _{0}^{t} f_\infty ^{\mathrm{ISI}}(x) \mathrm{d}x \le \int _{0}^{t} f_0^{\mathrm{ISI}}(x)\mathrm{d}x = F_0^{\mathrm{ISI}}(t) \end{aligned} \end{aligned}$$
(28)

where \(F_\infty ^{1}, F_\infty ^{\mathrm{ISI}}, F_0^{\mathrm{ISI}}\) and \(F_0^{F}\) are the cumulative probability distributions of the inter-spike intervals and first spikes, respectively. This is illustrated in Fig. 3 showing that indeed the first spikes move forward through STDP and the later spikes are more separated, which is consistent with the results from previous sections.

It is worth noting that our results are only valid for the specific case where the plasticity rule potentiates the presynaptic spike to a neuron before its postsynaptic spikes and depresses those afterward. As there is a zoo of possible time-dependent rules, we performed a short overview of the effects of those rules in “Appendix B,” finding that in a wide range of variants our results still hold. Another important feature that we have to consider is the addition of recurrent connections, which we address in “Appendix C”.

Fig. 3
figure 3

Evolution of the spike train: We plot cumulative probability distribution of the time of the first spike, the inter-spike interval when a presynaptic spike train is presented for the first time (left) and after many repetitions (center) and the number of spikes per bins of \(4\,\)ms on the first \(50\,\)ms of a spike train (left). We simulate 2000 neurons each receiving a presynaptic spike train lasting \(600\,\)ms with 200 presynaptic spikes, both inhibitory and excitatory, and whose arrival time is uniformly sampled. Every synapse evolves through STDP and being subject to both the fixed spike train with probability 0.33 and a random pair of pre- and postsynaptic spikes with \(t_{\mathrm{post}} - t_{\mathrm{pre}} \in \left[ -20\,\text { ms}, 20\,\text {ms}\right] \) with probability 0.66. We plot the time of the first spike (blue) and the inter-spike interval for second, third and fourth spikes, but subtracting the refractory period to have a pertinent comparison with the first spiking time. We can see that initially the first spike time is the same as the inter-spike interval for all the spikes, but after STDP is applied the average time of the first spike reduces, implying that the blue line moves to the left with respect to the time before learning (in the black dotted line) while the average inter-spike intervals increase, thus moving the curves to the right. This changes the distribution of spikes to have more of them concentrated in the beginning of the spike train (color figure online)

For the next section, it will be convenient to look at the instantaneous firing rate, which is obtained by accumulating the times of all spikes.

$$\begin{aligned} s(t) = \lim \limits _{\Delta t\rightarrow 0} \sum _{k=1}^{\infty } \dfrac{\text {Pr}\left[ t_k\in \left[ t,t+\Delta t \right] \right] }{\Delta t} \end{aligned}$$
(29)

where \(t_k\) is the time of the kth spike. Since the time of the kth spike is the sum of the inter-spike intervals of the first \(k-1\) spikes and the first spike, and the probability of a sum is given by the convolution of the probability distributions, we can rewrite the previous function as

$$\begin{aligned} s(t)= & {} \left( f^1 + f^1*f^{\mathrm{ISI}} + f^1*f^{\mathrm{ISI}} *f^{\mathrm{ISI}} + \cdots \right) (t) \nonumber \\= & {} \left( f^1*\sum _{k=0}^{\infty }\left( f^{\mathrm{ISI}}\right) ^{*k} \right) \end{aligned}$$
(30)

where \(*\) is the convolution operator, \(^{*k}\) is the convolution power. Note that \(f^1\) and \(f^{\mathrm{ISI}}\) depend on how many times the input has been repeated. We will refer to the subindex 0 and \(\infty \) to refer, respectively, to the cases where the presynaptic spike train is presented for the first time or when it has been presented many times.

The postsynaptic spike trains generated by neural populations are instantiate codes that transmit information about presynaptic spikes to other neurons. As STDP is a learning mechanism that modifies the postsynaptic spike train, we expect that it should improve this encoding. Each input stimulus triggers spikes in a certain neural population, and every neuron in that population has a certain performance associated to it, the two most common performance measures being energy consumption and resistance to noise (Rappaport 1996).

If we take the number of postsynaptic spikes generated by the neural population as a proxy for the metabolic costs of encoding a stimulus, then we would expect that number to decrease as the stimulus is presented more often, so that the encoding of common stimuli incurs less metabolic costs.

To evaluate how the number of spikes evolves, we consider the evolution of the first spike and inter-spike-interval cumulative probability distributions from Fig. 3. On one hand, the fact that the first spike moves forward implies that there will be more spikes concentrated on a small region at the beginning, so if we consider a very short time interval the concentration of spikes will increase. However, as we increase the length of the stimulus, the average distance between spikes will start to depend mostly on the inter-spike interval, implying that the spike density will be lower. In more formal terms, the number of spikes is given by the integral

$$\begin{aligned} S= & {} \int _{0}^{T} s(t) \mathrm{d}t = \int _{0}^{T} f^1(t)\mathrm{d}t \nonumber \\&+ \int _{0}^{T} f^1 *\sum _{k=1}^{\infty }\left( f^{\mathrm{ISI}}\right) ^{*k} \mathrm{d}t , \end{aligned}$$
(31)

which is dominated by the first term when T is small and by the second term when T is large. This can be quantified by the ratio in the decrease in spikes

$$\begin{aligned} \dfrac{S_\infty - S_0}{S_0}, \end{aligned}$$
(32)

where \(S_0\) is the number of spikes before STDP and \(S_\infty \) is the number afterward. Naturally, there are many parameters that affect the change in the number of spikes, in particular the length of the stimuli and the input rate or how often the input is presented with respect to other stimuli, which are shown in Fig. 4. In general, in short time intervals at most one spike would be present, thus the disappearance of second spikes induced by the depressive side of STDP does not play a role; at the same time, the spikes that would appear by the fluctuations in input weight, and which would simply disappear by the same process if STDP was not present, remain. Hence, in that case the number of spikes increases, while for long spike intervals the number of spikes decreases.

It is worth noticing that the reduction in the number of spikes that we observe in Fig. 4 does not correspond to the reduction in spike count that STDP induces in Poissonian spike trains. We tested this by checking how a Poissonian spike train with the same STDP parameters and the same weight distribution and input rate as in Fig. 3 changed, and we found that this leads to an increase of \(10 \%\) in the number of postsynaptic spikes because excitatory presynaptic spikes tend to induce postsynaptic spikes, thus the excitatory weights systematically increase.

Fig. 4
figure 4

Spike count evolution: We simulated spike trains of various lengths for different parameters of the input rate (IR) and the probability that the stimulus is presented (SP), and when it is not we induce a random pair of pre-postsynaptic spikes in every synapse with a \(t_{\mathrm{post}} - t_{\mathrm{pre}} \in \left[ -20\,\text {ms}, 20\,\text {ms}\right] \). In either case we investigate the change in the number of spikes. As we can see, for long spike trains the inter-spike-intervals increase and thus the number of spikes decreases. For short spike trains, on the other hand, there is at most one spike that can fit, so the inter-spike intervals are irrelevant. Furthermore, the spikes in such short intervals are self-maintained when STDP is present: If a spike appears and disappears when the presynaptic weights evolve randomly, the presence of a postsynaptic spike will potentiate those weights, hence the spike will be maintained, implying that STDP increases the number of spikes in short time intervals

Table 2 Synchrony evolution: We simulated spike trains with different values of the presynaptic input rate (IR) and the probability that the stimulus is presented compared to random pair of spikes per synapse (SP), and then measured the change in \(\phi \) taking a time window of 100ms and using 1000 neurons. As we can see, the synchronization always increases

Besides the number of spikes, it is also interesting to note how the distribution of those spikes change. Specifically, as the first spikes move forward, the spike train will become more synchronous as the distribution of spiking times becomes sharper, as we can see in Fig. 3, where the postsynaptic spike train has a peak of spikes that grows after STDP is applied. We quantify this by counting the highest concentration of spikes in a small time window of size L with respect to the total number of spikes, which can be written as

$$\begin{aligned} \phi = \dfrac{\max _t\int _{t}^{t+L}s(t)\mathrm{d}t}{S}\dfrac{T}{L}, \end{aligned}$$
(33)

where T is the time interval for the full stimulus such that \(\frac{S}{T}\) is the average spike rate and \(\frac{\max _t\int \nolimits _{t}^{t+L}s(t)\mathrm{d}t}{L}\) is the highest rate in a time window of length L. For a random spike train, the highest rate of spikes in a time window of length L would be similar to the average firing rate, corresponding to a \(\phi \approx 1\). However, if many spikes concentrate in a small time window, the spike trains are synchronized and we obtain a high value of \(\phi \). The results of simulations for various parameters are presented in Table 2, where the increase in \(\phi \) can be easily seen.

Fig. 5
figure 5

Encoding predictions: schema for the emergence of predictions (left) and firing latencies of neurons in encoding population (right): An external event creates three stimulus that trigger all the neurons in corresponding distinct neural populations \(P1, \ P2,\ P3\) with the stimuli inducing spikes during the intervals \([0\,\text {ms},500\,\text {ms}]\) for P1, \([500\,\text {ms},1000\,\text {ms}]\) for P2 and \([1000\,\text {ms},1500\,\text {ms}]\) for P3, respectively. The three populations, with \(N=50\) neurons each, also have synapses between them with delays sampled from a uniform distribution between \(d_{PiPj} \in [1\,\text {ms},5\,\text {ms}]\). Originally, almost all neurons in each population fire only after receiving inputs from their respective stimuli, but after the external event is repeated very often, the inter-population connections become strong enough to trigger some spikes before the stimulus is received (color figure online)

5 The emergence of predictions

When a group of neurons encodes a stimulus we mean that those neurons fire when the stimulus is presented. However, the neurons themselves are not aware agents and do not know anything about that stimulus; they simply receive a spike train that is strong enough to trigger their spiking. From the point of view of an encoding neurons, there is no difference between the stimulus-induced presynaptic spike train and any other input spike train that always precedes the stimulus.

Combining this observation with the results from previous sections showing that neurons will fire at the onset of a frequent input spike train, we can conclude that a neuron that “encodes” a stimulus can start firing before the stimulus is presented if another stimulus appears before it. As an illustrative example, imagine listening to a melody. Different parts of the melody trigger the activity of different groups of neurons in the same area of the brain. If the melody is repeated very often, the neurons P1 that react to an early part of the melody will systematically fire before the neurons P2 that react to a later part. As the melody is repeated, neurons in P2 will always fire after receiving spikes from neurons in P1 and thus the synapses from P1 to P2 will be reinforced. Eventually, the reinforced synapses might trigger spikes in P2 before the late part of the melody sounds. This can be extended to more populations encoding more stimuli, and thus the whole melody is encoded through simultaneous activity of all the neurons which originally encode only separate notes. This is illustrated and simulated in Fig. 5.

It is important to notice here that the predictions that we mention here are restricted to stimuli sequences that can be identified from the first input, meaning that we are not addressing the case of two sequences of stimuli which start activating the same neural population and then go on to activate different populations. If we have two possible stimuli sequences which start equally, STDP would force some neurons associated to both possible sequences fire at the onset of the stimuli, meaning that the system would learn that both sequences might follow. However, the differentiation of the two sequences can only be done when the two diverge, so the system must learn to maintain memory traces of the stimuli, a process that can also be implemented by STDP with lateral inhibition (Klampfl and Maass 2013).

6 Discussion

In this paper, we start by analyzing and expanding previous findings on latency reduction (Song et al. 2000; Guyonneau et al. 2005). Then, we extend them to trains of spikes and show that those mechanisms lead to encoding the more common inputs with less spikes while concentrating the remaining spikes in smaller time windows. This leads us to the conclusion that STDP reduces the amount of spikes used to encode frequent stimuli, in line with the idea that metabolic efficiency is one of the guiding principles of the brain (Hasenstaub et al. 2010; Laughlin 2001). The same phenomena also synchronize spikes by concentrating them in small time windows. Following the idea that synchronization improves communication between neuronal assemblies (Singer 2011; Fries 2005; Von Der Malsburg 1994), the fact that synchronization is induced by STDP indicates that its effects can be interpreted in communication terms. Finally, we show that the latency reduction can explain how the nervous system learns to forecast even without any feedback.

This study is another example of how simple, well-known plasticity rules that are present at synaptic level lead to modifications that are advantageous at the organism level. Furthermore, the fact that the same mechanism improves the neural code and creates predictions might explain how the ability of the brain to make predictions—which is one of the core problems in cognitive science—could have emerged as a consequence of evolutionary pressures on metabolic cost and information transmission.

Naturally, our work is also interesting for researchers in machine learning, as it shows that Hebbian learning rules, which are classically used to infer or reinforce correlations (Dayan and Abbott 2001), can be used to find predictions by adding a temporal asymmetry in the synaptic plasticity kernel. Furthermore, the fact that the same mechanism gives rise to predictions and coding efficiency is another example of the intimate relationship between machine learning and coding (MacKay et al. 2003), thus it might be interesting for information theorists.

The results exposed here also open new questions. The effects of latency reduction in networks of neurons—in particular recurrent ones—or the potential use of this prediction capabilities of STDP for machine learning require further study but could be useful extensions. However, the most immediate question is whether this unsupervised process is used in the nervous system. An experimental study should identify the neurons that encode two temporally correlated stimuli and follow the evolution of latencies as the stimuli are repeated, while simultaneously ensuring that this process was due to STDP alone without interference of reward systems that have been previously proposed.