1 Introduction

Variability of temperature time series has been described in two ways. Some people concentrate on scaling behavior of correlations and its seeming power law behaviour (long range correlations). Popular methods are detrended fluctuation analysis (DFA) (Fraedrich and Blender 2003), and multifractal versions (Moon et al. 2018), power spectral densities (Fredriksen and Rypdal 2016) and wavelet analysis (Lovejoy 2018). Such results consider the complete power of climate fluctuations. They are useful for estimations of prediction errors (Massah and Kantz 2016; Ludescher et al. 2016) and the validation of more complicated general circulation models (Govindan et al. 2002).

Other projects have tried to identify characteristic timescales in the data (Ghil et al. 2002). Here people use for example singular spectrum analysis (SSA) (Plaut et al. 1995), versions of Monte Carlo SSA (Paluš and Novotná 2004), and again wavelet methods (Baliunas et al. 1997). Results of these methods also help to validate general circulation models and contribute to the physical understanding of the atmosphere as well as to make predictions if the component is significant like the El Nino–Southern Oscillation (Chekroun et al. 2011). They can explain observed slowdown (or fastening) of global warming (Steinman et al. 2015; Hu and Fedorov 2017).

The disadvantage of the first approach is that it neglects properties that are already known and reduces the information to one property which is difficult to interpret (Maraun et al. 2004). The disadvantage of the second approach is that it does not take the full power of the fluctuations into account (Lovejoy 2015).

While parts of the literature concentrating on characteristic timescales completely ignore the findings of power law scaling, others mention the so called continuum variability (Huybers and Curry 2006) and treat it as one of the features contained in temperature fluctuations. Recently interest in this continuum variability is growing again (Lovejoy 2018; Fredriksen and Rypdal 2017). Lovejoy (2015) suggest to filter the known frequency modes, just like it is normally done with the seasonal cycle, in order to understand the nature of the continuum better. Our approach in fact follows this spirit. We will apply a novel method based on DFA that in contrast to traditional usage of DFA concentrates on the characteristic timescales. Therefore we show connections between the two approaches mentioned at the beginning. However, we conclude that all the power of the fluctuations can be explained by a superposition of few simple processes for the time range under consideration.

We want to concentrate on climate data from Europe where a large number of stations with several decades of daily recordings exists. For this region there is already a long history of investigations on both individual stations or grid data. Most importantly, people have repeatedly reported a 7–8 year cycle in various stations and other records (Plaut et al. 1995; Paluš and Novotná 1998; Sen and Ogrin 2016), the origin of which is still unclear. Spacial patterns were described in Grieser et al. (2002), Pišoft et al. (2009) and Jajcay et al. (2016), however, to our knowledge there has not been an algorithmic investigation on a similar number of stations as we perform it.

We present the foundations of our method in Sect. 3, summarizing previous work in Meyer and Kantz (2019). In Sect. 4 we postulate our model for weather and macroweather variability in Europe and extend our method to processes with more than one characteristic timescale. In Sect. 5 we show results of a large scale algorithmic application of our method to station data of European temperatures. We are able to reproduce the 7–8 year cycle in part of this data. We show that our method is able to separate data with and without this frequency mode.

2 Data

We analyze station data of temperature and pressure in Europe (Tank et al. 2002). It can be downloaded from http://www.ecad.com. The resolution of the time series is one day. The datasets strongly vary in length. We restrict ourselves to long time series as described in Appendix 1.

3 The method

Fig. 1
figure 1

Left: pressure (top) and temperature data (bottom) from Potsdam station. The measured values are displayed with solid lines in the background. The anomalies are the dots in the foreground. Center: fluctuation function of the pressure anomalies (dots) fitted by an AR(1) model (solid line). The fit indicates a relaxation time of 3 days. Right: fluctuation function of the temperature anomalies (red) fitted by a scaling law of \(\alpha =0.65\), and the fluctuation function of the real temperature data (blue) fitted by our model (Eq. (3))

Our method is based on detrended fluctuation analysis (DFA) (Peng et al. 1994) (see Appendix 2 for implementation) and its theoretical understanding (Hoell and Kantz 2015b). DFA produces the so called fluctuation function F of the time series, which is in fact a nonlinear transformation of the autocorrelation function C

$$\begin{aligned} {F}_x^2(s)=\sigma _x^2\left( L_q(0,s)+2\sum _{t=1}^{s-1}C_{xx}(t)L_q(t,s)\right) . \end{aligned}$$
(1)

Here \(\sigma _x^2\) is the variance of the time series and \(L_q\) is a kernel described in Appendix 3. It turns out, that looking at F(s) has two advantages over looking at C(t): on the one hand it is numerically more stable and on the other hand DFA is able to remove polynomial trends from the signal, so they do not contribute to the fluctuation function (Höll et al. 2019). We only use DFA1, because the trend due to climate change is sufficiently small compared to the fluctuations (Massah and Kantz 2016).

Meyer and Kantz (2019) have shown that this method has the ability to uncover characteristic timescales in time series. The idea is to fit the theoretical fluctuation function of autoregressive models of different orders: AR(1) models for relaxation times, and AR(2) for noisy oscillations. The method works well for data that is dominated by one characteristic timescale, like approximations to yearly global mean temperatures (Meyer et al. 2018) or an 11-month averaged El Nino signal. We show the theoretical fluctuation functions in (16) and (23) in Appendix 3.

As a first example we want to show the application of the method to the time series of air pressure measured at Potsdam, Germany. We deseasonalize both the pressure and its variance in order to get a homogeneous time series. Therefore we calculate the long time climatological average pressure for each calender day and subtract it from each day. We also divide each value by the average variance for the calender day.

In the long time limit F(s) scales like 1/2, implying short range correlations. Figure 1 shows that the data can be described by an AR(1) model in a fairly good approximation. Obviously, for a complex system like the atmosphere, we would not claim that pressure is purely autoregressive and no other effects happen at characteristic timescales. However, we can see that those other timescales are not significant for a smoothed measure such as the DFA-fluctuation function. The method shows us a data model that can be used for stochastic predictions and qualitative understanding of the dynamics.

In Fig. 1 we also show temperatures and temperature anomalies (T with climatological average temperatures for each calender day subtracted) from Potsdam. In DFA the temperature anomalies exhibit anomalous scaling proportional to \(s^{0.65}\), which could be described by a model with long range correlations (Massah and Kantz 2016). In this article we want to show a different approach. Instead of looking at the anomalies, we look at \(F_T\) itself. We want to identify characteristic timescales and therefore extend our understanding beyond the description of emergent scaling. If the time series is indeed a superposition of several independent processes we can make use of the superposition principle of DFA (Hu et al. 2001)

$$\begin{aligned} {F}_{x+y}^2(s)={F}_x^2(s)+{F}_y^2(s). \end{aligned}$$
(2)

This will be important for our purpose in the next section.

4 Decomposition of one time series

Fig. 2
figure 2

Decomposition of the fluctuation function of Potsdam temperatures. The orange curves represent fits to the fluctuation functions. The orange circles mark the points for which the fit is performed. Left: short time behaviour \(F^2_X\) fitted to the full fluctuation function \(F_T\) (green ‘\(\curlywedge \)’), which yields a relaxation time of 3.8 days. The red ‘+’ denotes \(F^2_{T-X}=F^2_T-F^2_X\); center: seasonal cycle \(F^2_Y\), fitted to \(F^2_{T-X}\). The resulting \(F^2_{T-X-Y}\) is plotted as blue ‘x’; right: interseasonal variability \(F^2_Z\) fitted to \(F^2_{T-X-Y}\)

In the examples above fitting fluctuation functions was exclusively applied to systems with one dominant characteristic timescale. What are we supposed to do if the fluctuation function implies a more complex dynamics? In this section, by analysing historic temperature data from Potsdam station, we show that our method still works for systems with more than one characteristic timescale if these timescales are sufficiently well separated.

We assume that the time series can be written as a superposition of three processes

$$\begin{aligned} T_t\approx X_t+Y_t+Z_t, \end{aligned}$$
(3)

namely, a short range part \(X_t\), the seasonal cycle \(Y_t\) and one unspecified component, which describes interseasonal variability. We will see that these components are clearly separated in time, such that \(Y_t\) and \(Z_t\) might as well be interpreted as additional components of the noise input \(\epsilon _t\) of \(X_t\). We will describe \(X_t\) and \(Z_t\) by simple autoregressive models, while the seasonal cycle \(Y_t\) is a regular oscillation.

Our model contains some simplifications. This is not only because it only consists of three processes described by few parameters, but also our assumption of independence of the three processes. This is not always true as shown in Paluš (2014) and Jajcay et al. (2016). However our focus in this text are not the cross-scale interactions, but only the observation of the different timescales. We also do not discuss spatial correlations.

Temperature correlations for short times decay exponentially (Massah and Kantz 2016). The behaviour is apparent in Fig. 2 in the fluctuation function \(F_T(s)\) for small s. Other subseasonal timescales are not pronounced in the fluctuation function. This lack of characteristic timescales is known and the reason why predictions on these timescale are so difficult (Ghil et al. 2019). Hence \(X_t\) is best described by an autoregressive model of order one AR(1)

$$\begin{aligned} X_{t_X}=cX_{t_X-1}+\epsilon _{t_X} \quad \text{ and } \quad r=-1/\log (c), \end{aligned}$$
(4)

where the noise term \(\epsilon _{t_X}\) is the output of some process on a much smaller timescale. The relaxation time r is determined by the AR-parameter c. The shape of the distribution of \(\epsilon _{t_X}\) is not important for our model (Meyer and Kantz 2019), however, it seems to be well approximated by a Gaussian distribution. The variance \(\sigma ^2_\epsilon \) is a free parameter. For simplicity we will later look at the standard deviation \(\sigma _X\) of the process instead of \(\sigma _\epsilon \). They are related by Eq. (18).

Due to the temporal separation of the three processes \(X_t\), \(Y_t\) and \(Z_t\), and because autoregressive models are low pass filters, the impact of the slower processes can be neglected for short times. For Potsdam the fit of Eq. (16) to \(F_T\) for small s (see Fig. 2) yields a relaxation time of 3.8 days. We subtract the fitted AR(1) fluctuation function \(F^2_X\) from \(F^2_T\).

Since we now covered the fluctuations for short times we change the timescale to 10 days by dividing \(F_{T-X}\) and s by 10 and remove small values of s. The by far most pronounced mode in the fluctuation function is the seasonal cycle (Deng et al. 2018). Although it is not perfectly periodic (Paluš et al. 2005) it has a phase-locked frequency of 1/365.25 days, which is enforced by the periodic driving of the sun. We do not remove the cycle by subtracting the climatological average since this popular procedure does not account for the whole dynamics on the yearly timescale. It does not take into account that the phase of the oscillation fluctuates. Instead we approximate the fluctuation function \(F_Y\) by an AR(2) process (see Eq. (6)) with \(b=-1\), which is equivalent to a sinusoidal signal

$$\begin{aligned} Y_{t_Y}=\frac{\sigma _Y}{2} \sin \left( \frac{2\pi t_Y}{1 \mathrm {year}}\right) , \end{aligned}$$
(5)

which does not describe the phase fluctuations, but accounts for their full power. Figure 2 shows that this is a good approximation to the remaining fluctuation function. Again we subtract \(F^2_Y\) from \(F^2_{T-X}\), change the timescale to half a year (182.625 days) by dividing both s and F by 18.2625 and remove points with small s.

So what is left if the first two components are removed from the fluctuation function? Plaut et al. (1995) claim to have found an 7–8 year cycle in the record of central England temperatures. Since then this mode was found in several other time series of European temperatures (Paluš and Novotná 1998; Pišoft et al. 2009; Sen and Ogrin 2016). We want to model interseasonal fluctuations with an AR(2) model

$$\begin{aligned} Z_{t_Z}=aZ_{t_Z-1}+bZ_{t_Z-2}+\eta _{t_Z}, \end{aligned}$$
(6)

that—dependent on the parameters a, b—can exhibit oscillatory behaviour with period \(\tau \) (see Eq. (21)) or pure relaxations. Again we treat the standard deviation \(\sigma _{Z_t}\) of the process as a free parameter, which determines \(\sigma _\eta \) of the noise \(\eta \) according to Eq. (22)

For Potsdam and other stations we looked at, our fit seems to show good agreement with \(F^2_{T-X-Y}\), although the fluctuation function is not as smooth as for shorter times. We obtain a period of 7.5 years. The complete fluctuation function \(F^2_{X+Y+Z}\) is shown in Fig. 1(right panel). It shows excellent agreement with the measured \(F^2_T\).

It has been suggested that climate can be described by a continuum variability with anomalous scaling (Lovejoy 2015) and quasi-periodic perturbations. For the time-range under consideration, which is admittedly shorter than in most studies that concentrate on this issue, we do not see any indicators that such a model would be necessary. The scale invariance of the anomalies visible in Fig. 1(right) is explained in our model by superposition of white noise driving at each identified timescale. This is because AR(2) is a low pass filter which ’hides’ the fluctuations of its driver for short times. Several AR(2) processes with different periods therefore lead to a successive increase of the background noise for longer time. This emergent scaling is an alternative interpretation to dynamical long range correlations, the origin of which is still not fully understood (Fraedrich et al. 2004; Fredriksen and Rypdal 2017; Meyer and Kantz 2017).

5 Analysis for European station data

Fig. 3
figure 3

Left: color plot of relaxation times \(r_T\), obtained from the AR(1) parameter (4) for each station. The color map can be found in the histogram in the CENTER, where it is compared with the histogram of the results of the same procedure for pressure data \(r_p\) (green line). The distributions of \(r_T\) and \(r_p\) are similar to each other; right: \(r_T\) and \(r_p\) are plotted against each other (green ‘+’) for stations where pressure data is available. The plot shows, that there is no visible correlation between both. The purple heat map in the background represents values of \(r_T\) for all stations

Fig. 4
figure 4

The standard deviation in all three processes X, Y and Z represented by the parameters \(\sigma _X\) (left), \(\sigma _Y\) (center), and \(\sigma _Z\) (right). For \(\sigma _Z\) stations where no oscillatory behaviour in the range of 2.5–30 years was found are neglected and marked by black crosses. Top: results for all stations. Bottom: histograms with identical color-coding

The decomposition of the fluctuation function as described in the previous section was incorporated into an algorithm and applied to temperature data of stations around Europe. We obtain six parameters, r, \(\sigma _X\), \(\sigma _Y\), a, b, and \(\sigma _Z\), for each station, which model the fluctuations of timescales up to a bit more than a decade, describing weather and macroweather. The aim is to investigate the spatial dependence of these parameters and especially for which regions the AR(2) model (6) indicates an oscillatory mode on the timescale of several years. The period time \(\tau \) of this oscillation can be calculated from the parameters a and b.

The short time dynamics is described by the relaxation time r. We obtain it by fitting \(F_X\) to \(F_T\). The results are displayed in Fig. 3. r shows clear regional patterns, which are much stronger than the uncertainty of our method. The correlation time is short in the west and the north of Europe close to the Atlantic. It is longer in central and eastern Europe and around the Baltic sea. For most stations r is just a bit smaller than 4 days.

The dynamics of pressure data for short times turns out to be similar to that of temperature. Both describe the typical timescale of weather patterns. However, there seems to be no connection between \(r_T\) and \(r_p\) for one specific station. On the contrary the regional patterns of pressure strongly differ from those of temperature. Pressure values relax faster in the south compared to the north.

The variance of the noise \(\epsilon \) in the short time dynamics is very low close to the coasts and has its highest values in Russia and Scandinavia. The variances of all three processes X, Y and Z are displayed in Fig. 4. The amplitude \(\sigma _Y\) of seasonality is strongly correlated with \(\sigma _X\).

Fig. 5
figure 5

Left: results of period times \(\tau \) for all stations. Fitting parameters that do not yield a period or a period far off the investigated timescale are marked as black crosses. Periods between 10 and 22 times 6 months are marked with colored dots. Other periods between 5 and 30 times 6 months are marked with grey triangles. Center: histogram of colored and grey points. A Gaussian curve is fitted to the colored part via the method of least squares, yielding \(7.6\pm 1.8\) years as the average period time. Right: Fourier transform of correlations function averaged over all stations in \(\varOmega _+\) with 5–11 year period (red) vs Fourier transform of correlations function averaged over all other stations in \(\varOmega _-\) (black)

When fitting \(F_Z\) of the AR(2) model to \(F_{T-X-Y}\) we only distinguish two cases. Either the model has an oscillatory mode with a period \(\tau \) between 10 and 22 half years—or not. The set \(\varOmega _-\) of stations without this mode consists of three subsets:

  • if the slope is \(\propto s^{0.5}\) or ever more flat the algorithm indicates an oscillation with a very short period

  • if the slope is very steep (\(\propto s^{1.5}\)) the algorithm indicates an oscillation with a very long period

  • for stations in between the AR(2) model has real roots indicating no oscillation.

All three cases have in common, that there is no sharp crossover in the fitted s-range and therefore statements about periods can only be made with high uncertainty. Examples are shown in Appendix 1.

For 51% of the analyzed stations we observe an oscillation, see Fig. 5. We denote the set of these stations by \(\varOmega _+\). The categorization into the two subgroups uncovers regional patterns with few exceptions which might be interpreted as false classifications. The frequency mode is significant in an area including England, Belgium, the south of Scandinavia, central Europe north of the Alps and parts of eastern Europe. The values for \(\tau \) that we obtain do not show clear regional patterns, which indicates that we indeed only see the known 7–8 year cycle and the deviations are the uncertainty of our method. On average the period length is 7.6 years with large standard deviation of 1.8 years.

We test the ability of our method to distinguish between stations with and without an observable period by looking at the conditionally averaged power spectrum

$$\begin{aligned} P(\omega )={\mathcal {F}}[\langle C(t)\rangle _\varOmega ], \end{aligned}$$
(7)

where \({\mathcal {F}}\) is the Fourier transform and the brackets \(\langle \rangle \) denote the average over all stations in the set \(\varOmega _+\) or the average over all stations in the set \(\varOmega _-\). The validations is successful as the stations where the period was found show a clear peak at 8 years while the others do not (see Fig. 5 right).

6 Conclusions

We introduce a method for the detection of characteristic timescales in time series based on detrended fluctuation analysis that works for data where the dominant timescales are sufficiently well separated. It is an intuitive tool for the derivation of approximate dynamics that accounts for the full power of the fluctuations in the system. In this way we bridge the gap between spectral methods that are applied for the detection of oscillations and scaling methods that detect colored noise. The method does not detect nonlinear effects like cross-scale interactions but only covers the characteristic timescales individually.

While pressure data usually only shows exponential decay of correlations, temperature data shows a much slower decay, often interpreted as long range correlations. We model these time series with a simple model, a superposition of short range decay, a seasonal cycle and a potentially oscillating noisy component for longer times. This model is able to describe the observed fluctuation functions accurately. DFA is a smoothing method that is not designed to find every tiny structure in the time series. For timescales up to a decade it is not necessary to claim a colored background noise, as many authors do, who look at longer timescales. The increase of the background noise can be explained by the input noise of the interseasonal variability. By fitting the parameters we obtain the power of each component for each station. more importantly we are able to reproduce a previously found 7–8 period. Our method is with only few false classifications able to distinguish between stations where this component is significant, and stations, where it is not.