Abstract
Many atmospheric fields—in particular the temperature—respect statistical symmetries that characterize the macroweather regime, i.e. time-scales between the \(\approx\) 10 day lifetime of planetary sized structures and the (currently) 10–20 year scale at which the anthropogenic forcings begin to dominate the natural variability. The scale-invariance and the low intermittency of the fluctuations implies the existence of a huge memory in the system that can be exploited for macroweather forecasts using well-established (Gaussian) techniques. The Stochastic Seasonal to Interannual Prediction System (StocSIPS) is a stochastic model that exploits these symmetries to perform long-term forecasts. StocSIPS includes the previous ScaLIng Macroweather Model (SLIMM) as a core model for the prediction of the natural variability component of the temperature field. Here we present the theory for improving SLIMM using discrete-in-time fractional Gaussian noise processes to obtain an optimal predictor as a linear combination of past data. We apply StocSIPS to the prediction of globally-averaged temperature and confirm the applicability of the model with statistical testing of the hypothesis and a good agreement between the hindcast skill scores and the theoretical predictions. Finally, we compare StocSIPS with the Canadian Seasonal to Interannual Prediction System. From a forecast point of view, GCMs can be seen as an initial value problem for generating many “stochastic” realizations of the state of the atmosphere, while StocSIPS is effectively a past value problem that estimates the most probable future state from long series of past data. The results validate StocSIPS as a good alternative and a complementary approach to conventional numerical models. Temperature forecasts using StocSIPS are published on a regular basis in the website: http://www.physics.mcgill.ca/StocSIPS/.
Similar content being viewed by others
References
Baillie RT, Chung S-K (2002) Modeling and forecasting from trend-stationary long memory models with applications to climatology. Int J Forecast 18:215–226. https://doi.org/10.1016/S0169-2070(01)00154-6
Berkeley Earth (2018) Land + Ocean (1850–Recent). http://berkeleyearth.org/data/. Accessed 21 May 2018
Berner J, Achatz U, Batté L et al (2017) Stochastic parameterization: toward a new view of weather and climate models. Bull Am Meteorol Soc 98:565–588. https://doi.org/10.1175/BAMS-D-15-00268.1
Biagini F, Hu Y, Øksendal B, Zhang T (2008) Stochastic calculus for fractional brownian motion and applications. Springer, London
Blender R, Fraedrich K, Hunt B (2006) Millennial climate variability: GCM-simulation and Greenland ice cores. Geophys Res Lett 33:L04710. https://doi.org/10.1029/2005GL024919
CanSIPS (2016) CanSIPS data in GRIB2 format. https://weather.gc.ca/grib/grib2_cansips_e.html. Accessed 16 Feb 2016
Cowtan K, Way RG (2014) Coverage bias in the HadCRUT4 temperature series and its impact on recent temperature trends. Q J R Meteorol Soc 140:1935–1944. https://doi.org/10.1002/qj.2297
Cowtan and Way (2018) Coverage bias in the HadCRUT4 temperature record. http://www-users.york.ac.uk/~kdc3/papers/coverage2013/series.html. Accessed 21 May 2018
Crochemore L, Ramos M-H, Pappenberger F (2016) Bias correcting precipitation forecasts to improve the skill of seasonal streamflow forecasts. Hydrol Earth Syst Sci 20:3601–3618. https://doi.org/10.5194/hess-20-3601-2016
Franzke C (2012) Nonlinear trends, long-range dependence, and climate noise properties of surface temperature. J Clim 25:4172–4183. https://doi.org/10.1175/JCLI-D-11-00293.1
Franzke CLE, O’Kane TJ, Berner J et al (2014) Stochastic climate theory and modeling. Wiley Interdiscip Rev Clim Chang 6:63–78. https://doi.org/10.1002/wcc.318
Garcia LE, Dawdy DR, Mejia JM (1972) Long memory monthly streamflow simulation by a broken line model. Water Resour Res 8:1100–1105. https://doi.org/10.1029/WR008i004p01100
GISTEMP Team (2018) GISS Surface Temperature Analysis (GISTEMP). NASA Goddard Institute for Space Studies. https://data.giss.nasa.gov/gistemp/. Accessed 21 May 2018
Gleason B, Williams C, Menne M, Lawrimore J (2015) GHCN-M Technical Report No. GHCNM-15-01 Modifications to GHCN-Monthly (version 3.3.0) and USHCN (version 2.5.5) processing systems. Asheville, NC
Gneiting T, Raftery AE, Westveld AH, Goldman T (2005) Calibrated probabilistic forecasting using ensemble model output statistics and minimum CRPS estimation. Mon Weather Rev 133:1098–1118. https://doi.org/10.1175/MWR2904.1
Gripenberg G, Norros I (1996) On the prediction of fractional brownian motion. J Appl Probab 33:400–410. https://doi.org/10.2307/3215063
Hansen J, Ruedy R, Sato M, Lo K (2010) Global surface temperature change. Rev Geophys 48:RG4004. https://doi.org/10.1029/2010rg000345
Hasselmann K (1976) Stochastic climate models Part I. Theory. Tellus 28:473–485. https://doi.org/10.1111/j.2153-3490.1976.tb00696.x
Hébert R, Lovejoy S, Tremblay B (2019) An observation-based scaling model for climate sensitivity estimates and global projections to 2100. Clim Dyn (under review)
Hersbach H (2000) Decomposition of the continuous ranked probability score for ensemble prediction systems. Weather Forecast 15:559–570. https://doi.org/10.1175/1520-0434(2000)015%3c0559:DOTCRP%3e2.0.CO;2
Hipel KW, McLeod AI (1994) Time series modelling of water resources and environmental systems, 1st edn. Elsevier, Amsterdam
Huybers P, Curry W (2006) Links between annual, Milankovitch and continuum temperature variability. Nature 441:329
Jeong H-DJ, Pawlikowski K, McNickle DC (2003) Generation of self-similar processes for simulation studies of telecommunication networks. Math Comput Model 38:1249–1257. https://doi.org/10.1016/S0895-7177(03)90127-0
Keller JD, Hense A (2011) A new non-Gaussian evaluation method for ensemble forecasts based on analysis rank histograms. Meteorol Zeitschrift 20:107–117. https://doi.org/10.1127/0941-2948/2011/0217
Kharin VV, Zwiers FW (2003) Improved seasonal probability forecasts. J Clim 16:1684–1701. https://doi.org/10.1175/1520-0442(2003)016%3c1684:ISPF%3e2.0.CO;2
Kharin VV, Teng Q, Zwiers FW et al (2009) Skill assessment of seasonal hindcasts from the Canadian historical forecast project. Atmos Ocean 47:204–223. https://doi.org/10.3137/AO1101.2009
Kharin VV, Merryfield WJ, Boer GJ, Lee W-S (2017) A postprocessing method for seasonal forecasts using temporally and spatially smoothed statistics. Mon Weather Rev 145:3545–3561. https://doi.org/10.1175/MWR-D-16-0337.1
Koscielny-Bunde E, Bunde A, Havlin S et al (1998) Indication of a universal persistence law governing atmospheric variability. Phys Rev Lett 81:729–732. https://doi.org/10.1103/PhysRevLett.81.729
Lettenmaier DP, Burges SJ (1977) Operational assessment of hydrologic models of long-term persistence. Water Resour Res 13:113–124. https://doi.org/10.1029/WR013i001p00113
Lorenz EN (1963) Deterministic nonperiodic flow. J Atmos Sci 20:130–141. https://doi.org/10.1175/1520-0469(1963)020%3c0130:DNF%3e2.0.CO;2
Lorenz EN (1972) Predictability; does the flap of a butterfly’s wings in Brazil set off a tornado in Texas? Presented before the American Association for the Advancement of Science, December 29, 1972. http://eaps4.mit.edu/research/Lorenz/Butterfly_1972.pdf
Lovejoy S (2014) Scaling fluctuation analysis and statistical hypothesis testing of anthropogenic warming. Clim Dyn 42:2339–2351. https://doi.org/10.1007/s00382-014-2128-2
Lovejoy S (2015) Using scaling for macroweather forecasting including the pause. Geophys Res Lett 42(17):7148–7155. https://doi.org/10.1002/2015GL065665
Lovejoy S (2017) How accurately do we know the temperature of the surface of the earth? Clim Dyn 49:4089–4106. https://doi.org/10.1007/s00382-017-3561-9
Lovejoy S, de Lima MIP (2015) The joint space-time statistics of macroweather precipitation, space-time statistical factorization and macroweather models. Chaos An Interdiscip J Nonlinear Sci 25:075410. https://doi.org/10.1063/1.4927223
Lovejoy S, Schertzer D (2012) Low-frequency weather and the emergence of the climate. Extrem Events Nat Hazards Complex Perspect 196:231–254
Lovejoy S, Schertzer D (2013) The weather and climate: emergent laws and multifractal cascades. Cambridge University Press, Cambridge
Lovejoy S, Schertzer D, Varon D (2013) Do GCMs predict the climate… or macroweather? Earth Syst Dyn 4:439–454. https://doi.org/10.5194/esd-4-439-2013
Lovejoy S, del Rio Amador L, Hébert R (2015) The ScaLIng Macroweather Model (SLIMM): using scaling to forecast global-scale macroweather from months to decades. Earth Syst Dyn 6:637–658. https://doi.org/10.5194/esd-6-637-2015
Lovejoy S, Del Rio Amador L, Hébert R (2018) Harnessing butterflies: theory and practice of the stochastic seasonal to interannual prediction system (StocSIPS) BT. In: Tsonis AA (ed) Advances in nonlinear geosciences. Springer, Cham, pp 305–355
Mandelbrot BB (1971) A fast fractional gaussian noise generator. Water Resour Res 7:543–553. https://doi.org/10.1029/WR007i003p00543
Mandelbrot BB (1972) Broken line process derived as an approximation to fractional noise. Water Resour Res 8:1354–1356. https://doi.org/10.1029/WR008i005p01354
Mandelbrot BB, Van Ness JW (1968) Fractional Brownian motions, fractional noises and applications. SIAM Rev 10:422–437. https://doi.org/10.1137/1010093
Mandelbrot BB, Wallis JR (1969) Computer Experiments with fractional gaussian noises: part 3, mathematical appendix. Water Resour Res 5:260–267. https://doi.org/10.1029/WR005i001p00260
Matalas NC, Wallis JR (1971) Statistical properties of multivariate fractional noise processes. Water Resour Res 7:1460–1468. https://doi.org/10.1029/WR007i006p01460
Meinshausen M, Smith SJ, Calvin K et al (2011) The RCP greenhouse gas concentrations and their extensions from 1765 to 2300. Clim Change 109:213–241. https://doi.org/10.1007/s10584-011-0156-z
Mejia JM, Rodriguez-Iturbe I, Dawdy DR (1972) Streamflow simulation: 2. The broken line process as a potential model for hydrologic simulation. Water Resour Res 8:931–941. https://doi.org/10.1029/WR008i004p00931
Merryfield WJ, Bertrand D, Fontecilla J-S et al (2011) The Canadian Seasonal to Interannual Prediction System (CanSIPS)—an overview of its design and operational implementation—technical Note
Merryfield WJ, Lee W-S, Boer GJ et al (2013) The Canadian Seasonal to Interannual Prediction System. Part I: models and initialization. Mon Weather Rev 141:2910–2945. https://doi.org/10.1175/MWR-D-12-00216.1
Met Office Hadley Centre (2018) Met Office Hadley Centre observations datasets. http://www.metoffice.gov.uk/hadobs/hadcrut4/data/current/download.html. Accessed 21 May 2018
Morice CP, Kennedy JJ, Rayner NA, Jones PD (2012) Quantifying uncertainties in global and regional temperature change using an ensemble of observational estimates: the HadCRUT4 data set. J Geophys Res Atmos. https://doi.org/10.1029/2011jd017187
Newman M, Sardeshmukh PD, Winkler CR, Whitaker JS (2003) A study of subseasonal predictability. Mon Weather Rev 131:1715–1732. https://doi.org/10.1175//2558.1
NOAA-NCEI (2018) Global Surface Temperature Anomalies. NOAA National Center for Environmental Information. https://www.ncdc.noaa.gov/monitoring-references/faq/anomalies.php. Accessed 21 May 2018
Norros I (1995) On the use of fractional Brownian motion in the theory of connectionless networks. IEEE J Sel Areas Commun 13:953–962. https://doi.org/10.1109/49.400651
O’Connell PE (1974) Stochastic modelling of long-term persistence in streamflow sequences. Doctoral Thesis. Imperial College, London
Palma W (2007) Long-memory time series. Wiley, Hoboken
Palmer T, Buizza R, Hagedorn R et al (2006) Ensemble prediction: a pedagogical perspective. ECMWF Newsl 2:10–17. https://doi.org/10.21957/ab129056ew
Papoulis A, Pillai SU (2002) Probability, random variables and stochastic processes, 4th edn. McGraw-Hill, New York
Pasternack A, Bhend J, Liniger MA et al (2018) Parametric decadal climate forecast recalibration (DeFoReSt 1.0). Geosci Model Dev 11:351–368. https://doi.org/10.5194/gmd-11-351-2018
Paxson V (1997) Fast, approximate synthesis of fractional Gaussian noise for generating self-similar network traffic. ACM SIGCOMM Comput Commun Rev 27:5–18. https://doi.org/10.1145/269790.269792
Penland C, Matrosova L (1994) A balance condition for stochastic numerical models with application to the El niño–Southern oscillation. J Clim 7:1352–1372. https://doi.org/10.1175/1520-0442(1994)007%3c1352:ABCFSN%3e2.0.CO;2
Penland C, Sardeshmukh PD (1995) The optimal growth of tropical sea surface temperature anomalies. J Clim 8:1999–2024. https://doi.org/10.1175/1520-0442(1995)008%3c1999:TOGOTS%3e2.0.CO;2
Rodriguez-Iturbe I, Mejia JM, Dawdy DR (1972) Streamflow Simulation: 1. A new look at Markovian Models, fractional Gaussian noise, and Crossing Theory. Water Resour Res 8:921–930. https://doi.org/10.1029/WR008i004p00921
Rohde R, Muller RA, Jacobsen R et al (2013) A new estimate of the average earth surface land temperature spanning 1753 to 2011. Geoinform Geostat Overv. https://doi.org/10.4172/2327-4581.1000101
Rypdal K, Østvand L, Rypdal M (2013) Long-range memory in Earth’s surface temperature on time scales from months to centuries. J Geophys Res Atmos 118:7046–7062. https://doi.org/10.1002/jgrd.50399
Sardeshmukh PD, Sura P (2009) Reconciling non-Gaussian climate statistics with linear dynamics. J Clim 22:1193–1207. https://doi.org/10.1175/2008JCLI2358.1
Smith TM, Reynolds RW, Peterson TC, Lawrimore J (2008) Improvements to NOAA’s historical merged land-ocean surface temperature analysis (1880–2006). J Clim 21:2283–2296. https://doi.org/10.1175/2007JCLI2100.1
Stanski HR, Wilson LJ, Burrows WR (1989) Survey of common verification methods in meteorology. WWW Tech. Rep. 8, WMO/TD 358. Ontario
Van Schaeybroeck B, Vannitsem S (2018) Postprocessing of long-range forecasts. Statistical postprocessing of ensemble forecasts. Elsevier, Oxford, pp 267–290
Winkler CR, Newman M, Sardeshmukh PD (2001) A linear model of wintertime low-frequency variability. Part I: formulation and forecast skill. J Clim 14:4474–4494. https://doi.org/10.1175/1520-0442(2001)014%3c4474:ALMOWL%3e2.0.CO;2
Wold H (1938) A study in the analysis of stationary time series. Almqvist und Wiksell, Uppsala
Yuan N, Fu Z, Liu S (2015) Extracting climate memory using fractional integrated statistical model: a new perspective on climate prediction. Sci Rep 4:6577. https://doi.org/10.1038/srep06577
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix 1: Simulation, parameters estimation, ergodicity and model adequacy
1.1 Simulation
When modeling real time series and testing numerical algorithms, it is often useful to obtain synthetic realizations of fGn processes. There are many methods for simulating approximate samples of fGn, e.g.: (1) type 1 (Mandelbrot and Wallis 1969), (2) type 2 (Mandelbrot and Wallis 1969), (3) fast fGn (Mandelbrot 1971), (4) filtered fGn (Matalas and Wallis 1971), (5) ARMA(1,1) (O’Connell 1974), (6) broken line (Garcia et al. 1972; Mejia et al. 1972; Rodriguez-Iturbe et al. 1972; Mandelbrot 1972), (7) ARMA-Markov models (Lettenmaier and Burges 1977) and some approximate, more efficient, recent methods (Paxson 1997; Jeong et al. 2003). We can choose among these methods based on their strengths and weaknesses, depending on the specific application we need.
Nevertheless, instead of using short memory approximations for simulating fGn, it is possible to generate exact realizations by applying the following procedure (Hipel and McLeod 1994; Palma 2007). In Eq. (20) we gave the MA representation of our series for any time, \(t\), based on the knowledge of an infinite past of innovations, \(\left\{ {\gamma_{t - j} } \right\}_{j = 1, \ldots ,\infty }\) with \(\gamma_{t} \sim {\text{NID}}\left( {0,1} \right)\) and \(\langle \gamma_{i} \gamma_{j} \rangle = \delta_{ij}\). If we want a series with specific length, \(N\), mean \(\mu\), variance \(\sigma_{T}^{2}\) and fluctuation exponent \(H\), we can work in a similar way as we did with the AR representation for obtaining the predictor. By replacing the coefficients, \(\varphi_{j}\), we could write instead the finite sum:
for \(t = 1, \ldots ,N\), where the optimal coefficients \(m_{ij}\) are the elements of the lower triangular matrix \({\mathbf{M}}_{{H,\sigma_{T} }}^{N}\) given by the Cholesky decomposition of the autocovariance matrix, \({\mathbf{R}}_{{H,\sigma_{T} }}^{N} = \left[ {C_{{H,\sigma_{T} }} \left( {i - j} \right)} \right]_{i,j = 1, \ldots ,N}\); that is:
with \(m_{ij} = 0\) for \(j > i\). In summary, for obtaining an fGn realization of length \(N\), we need to generate a white-noise process \(\left\{ {\gamma_{t} } \right\}_{t = 1, \ldots ,N}\) with an appropriate method, obtain the autocovariance matrix \({\mathbf{R}}_{{H,\sigma_{T} }}^{N}\) using Eq. (7.iii), then get \({\mathbf{M}}_{{H,\sigma_{T} }}^{N}\) from the Cholesky decomposition of \({\mathbf{R}}_{{H,\sigma_{T} }}^{N}\), and finally apply Eq. (46) for every \(t\) to obtain our \(\left\{ {T_{t} } \right\}\) series. The variables \(T_{t}\) will be \({\text{NID}}\left( {\mu ,\sigma_{T}^{2} } \right)\) and the process will have fluctuation exponent \(H\) in the interval \(\left( { - 1, 0} \right)\).
1.2 Maximum likelihood estimation
If instead of simulating an fGn process, we are interested in the opposite operation of finding the parameters that best fit a given time series, the most accurate method to use is based on maximizing the log-likelihood function (Hipel and McLeod 1994). Suppose that we have our vector \({\mathbf{T}}_{N} = \left[ {T_{1} , \ldots ,T_{N} } \right]^{T}\) that represents a stationary Gaussian process. Then the log-likelihood function of this process is given by:
where \({\tilde{\mathbf{T}}}_{N,\mu } = \left[ {T_{1} - \mu , \ldots ,T_{N} - \mu } \right]^{T}\) is a vector formed by our original series after removing the mean.
For fixed \(H\), the maximum likelihood estimators (MLE) of \(\mu\) and \(\sigma_{T}\) are:
and
where \({\mathbf{1}}_{N} = \left[ {1,1, \ldots ,1} \right]^{T}\) is an \(N \times 1\) vector with all the elements equal to 1 and \({\tilde{\mathbf{R}}}_{H}^{N} = {\mathbf{R}}_{{H,\sigma_{T} }}^{N} /\sigma_{T}^{2}\) is the autocorrelation matrix, which only depends on \(H\).
Substituting these values into Eq. (48), we obtain the maximized log-likelihood function of \(H\):
The estimate for the fluctuation exponent, \(\hat{H}_{l}\), is obtained by maximizing \({\mathcal{L}}_{ \text{max} } \left( H \right)\) and can be used then to obtain \(\hat{\mu }\) and \(\hat{\sigma }_{T}^{2}\) using Eqs. (49) and (50).
1.3 Ergodicity
It is worth noticing here that \(\hat{\mu }\) and \(\hat{\sigma }_{T}^{2}\) are estimates of the ensemble mean \(\mu = \left\langle {T_{t} } \right\rangle\) and variance \(\sigma_{T}^{2} = \langle {\left( {T_{t} - \mu } \right)^{2} } \rangle\) of the fGn process, respectively (see Sect. 2.1). If we try to estimate these parameters based on temporal averages of a single realization, some differences may arise with the values obtained using Eqs. (49) and (50). To explain these differences, we briefly discuss some ergodic properties of fGn processes.
Let
and
be the temporal average estimates of the mean and the variance of our process, respectively (the overbar indicates temporal averaging, \(N\) is considered large here), SD indicates “standard deviation”.
Using the relationship between fBm and fGn (Eq. (5)), we can write the temperature as:
The fBm process has the following properties:
Usually, the condition \(B_{{H^{\prime}}} \left( 0 \right) = 0\) is added to this definition. Using this and Eq. (54), by telescopic sum all addends cancel except for the last one and we obtain:
Taking ensemble averages and using Eqs. (55) (ii) and (iii) we get:
and
where we replaced \(H' = H + 1\).
Consequently, since the process \(B_{{H^{\prime}}} \left( t \right)\) is Gaussian, we conclude that, the temporal average estimate of the mean satisfies:
Now, taking the ensemble average of Eq. (53), we get:
The ensemble and the time averaging operations commute in the first term of the right-hand side of Eq. (60):
Using this and Eq. (58) for the last term in Eq. (60), we finally get:
meaning that the temporal average \(SD_{T}\) is a biased estimate of the variance of the process, \(\sigma_{T}^{2}\). An unbiased estimate would then be \(SD_{T}^{2} /\left( {1 - N^{2H} } \right)\). The variance of this estimator is more difficult to obtain. Its derivation, together with potential applications for treating climate series, will be presented in a future paper (currently in preparation).
In the limit \(N \to \infty\), as \(- 1 < H < 0\), we have \(SD_{T}^{2} \to \sigma_{T}^{2}\), meaning that the process is ergodic (the temporal average and the ensemble average coincide for infinitely long series). Nevertheless, for \(H \to 0\) this convergence is very slow, and a very long series would be needed in order to estimate the variance of the process from the sample variance without any correction. For example, for \(H = -\, 0.1\) and \(N = 1656\) months \(= 138\) years (realistic values for globally-averaged temperatures, see Sect. 3), we have \(SD_{T}^{2} /\sigma_{T}^{2} = \left( {1 - N^{2H} } \right) = 0.772\), i.e. a 23% difference between both estimates. In the same sense, if we want to estimate \(\sigma_{T}^{2}\) from the sample variance with 95% accuracy, we would need a series with \(N = 3.2 \cdot 10^{6}\) (if \(N\) is in months that would be \(N =\) 266,667 years!). The last three columns of Table 4 show the average estimates \(\hat{\sigma }_{T} = \sqrt {\hat{\sigma }_{T}^{2} }\) (Eq. (50)), \(SD_{T}\) (Eq. (53)) and the confirmation of their relationship (Eq. (62)), for simulations of fGn with length \(N = 1656\) and parameters \(\mu = 0\), \(\sigma_{T} = 1\) and values of \(H\) in the range \(\left( { -1/2, 0} \right)\). In each case, 200 realizations were analyzed, but only the average values of the estimates are shown. The standard deviations are always 2–7% of the respective mean values and were not reported. Notice that the difference between \(\hat{\sigma }_{T}\) and \(SD_{T}\) increases as \(H\) goes close to zero and the memory effects become more important.
Let us return now to the estimates \(\hat{\mu }\) and \(\hat{\sigma }_{T}^{2}\) given by Eqs. (49) and (50), respectively. These ensemble estimates are still obtained from the information of only one finite series, \({\mathbf{T}}_{N} = \left[ {T_{1} , \ldots ,T_{N} } \right]^{T}\), but the presence of the correlation matrix, \({\tilde{\mathbf{R}}}_{H}^{N}\), automatically includes all the information from the infinite unknown past. If we make \({\tilde{\mathbf{R}}}_{H}^{N} = {\mathbf{I}}_{\varvec{N}}\) (\({\mathbf{I}}_{N}\) is the \(N \times N\) identity matrix) in Eqs. (49) and (50) (or equivalently \(H = - 1/2\)), we obtain:
and
This means that the temporal average estimates based on one realization of the process are only valid for uncorrelated process, for which the ensemble and the sample averages are equal. When both correlations and memory effects are present, this information must be considered. In the case of fGn processes, the memory effects are introduced by including the correlation matrix which only depends on the fluctuation exponent \(H\). The value of this parameter for the process can also be obtained from only one realization of the same as shown below.
1.4 Quasi-maximum-likelihood estimation for \(H\)
As we mentioned before, the MLE for the fluctuation exponent, \(\hat{H}_{l}\), is obtained by maximizing \({\mathcal{L}}_{ \text{max} } \left( H \right)\) (Eq. (51)). The process of optimization of \({\mathcal{L}}_{ \text{max} } \left( H \right)\) could easily be computationally expensive for large values of \(N\). To avoid this, many approximate methods have been developed. We can use Eq. (9) to obtain \(\hat{H}_{s} = \left( {\beta_{l} - 1} \right)/2\) from the spectral exponent at low frequencies. This method, as well as the Haar wavelet analysis to obtain an estimate \(\hat{H}_{h}\) from the exponent of the Haar fluctuations, was used in Lovejoy and Schertzer (2013) and Lovejoy et al. (2015) to obtain estimates of \(H\) for average global and Northern Hemisphere anomalies. These two methods depend on the range selected for the linear regression and, when the graphs are noisy, it could result in poor estimates of the exponents. They, nevertheless, have the advantage of being more general; they yield \(H\) estimates even for highly nonGaussian processes. In the present case, a more accurate approximation is based on quasi-maximum-likelihood estimates (QMLE) from autoregressive approximations (Palma 2007).
Suppose we have a series of \(N\) observations, \(\left\{ {T_{t} } \right\}_{t = 1, \ldots ,N}\), we can build the one-step predictor for \(T_{t}\), \(\hat{T}_{t}^{p} \left( 1 \right)\) from Eq. (22) using a memory of \(p\) steps in the past with \(p + 1 < t \le N\):
Then, the approximate QMLE, \(\hat{H}_{q}\), is obtained by minimizing the function
Remember that the coefficients \(\phi_{p,j}\) only depend on \(H\). An added advantage of this method is that, by construction, it is done as part of the verification process based on hindcasts. The actual mean square error (MSE) of our one-step predictor with memory \(p\) is \({\mathcal{L}}_{1} \left( H \right)/\left( {N - p - 1} \right)\), so in practice, we perform the one-step hindcasts for different values of \(H\) in the specified range and select the value that gives the minimum MSE. The computation of the coefficients \(\phi_{pj}\) is fast, since we do not need to take very large values of \(p\) to achieve nearly the asymptotic skill, as we showed in Sect. 2.2.1.
In order to compare these different estimation methods, we performed some numerical experiments. By using Eq. (46) for the exact method with parameters \(\mu = 0\) and \(\sigma_{T} = 1\), we generated fGn ensembles of one hundred members of length \(N = 1656\) (see Sect. 3) for each value of \(H \in \left\{ { -\, 0.45, - \,0.40, -\, 0.35, -\, 0.30, -\, 0.25, - \,0.20, -\, 0.15, -\, 0.10, -\, 0.05} \right\}\). Then, we estimated \(H\) from the four previously mentioned methods for each realization. The results are summarized in Table 4. The values in parentheses represent the standard deviations for each ensemble. The maximum likelihood, the Haar fluctuation and the spectral methods allow for direct estimates of the ensemble values (shown with the subscript “ens” in Table 4) by considering the maximum likelihood of the vector process, the ensemble of all the fluctuations or the average of all the spectra, respectively from all the paths instead of from each of the series independently. We could say, for example that \(\hat{H}_{s}\) is the mean of all the \(\hat{H}_{s}\)’s obtained from each realization spectrum, while \(\hat{H}_{{s,{\text{ens}}}}\) is the value obtained from the mean of all the spectra. This ensemble estimate reduces the error due to dispersion of each of the ensemble members. For the QMLE, a memory \(p = 20\) was used.
As we can see from Table 4, for the MLE method, there is good agreement between the average of the estimates for each realization and the direct ensemble estimate. This is not the case for the less accurate methods of Haar fluctuation and spectral analysis in the member-by-member cases. Comparatively, the standard deviation of these two methods (without considering the estimation error for each specific realization) is much larger than for the MLE. Nevertheless, the ensemble estimates for the Haar are very accurate because the dispersion for the ensemble is much lower than for each individual graph. In practice, it is almost always the case that we only have a given time series to analyze instead of multiple realizations of an ensemble. In that sense, unless we have more theoretical or empirical justifications for the scaling, estimates based on these graphical methods should be considered cautiously.
A direct comparison of the second and third columns in Table 4 shows the accuracy of the QMLE method if we take the MLE as reference. The average values and the standard deviations for the two methods are very close for small values of \(H\), but as we move to values close to zero there is a systematic bias in the QMLE method towards slightly smaller values than those obtained with the MLE. Nevertheless, the presence of this bias is of little consequence from the point of view of forecasting and can be reduced by increasing the memory used. As we mentioned before, the QMLE method is based on minimizing the MSE—or what is the same—maximizing the MSSS obtained from hindcasts. Near the extreme, a small variation of the value of \(H\) used to perform the forecast will produce almost no change on the MSSS obtained.
1.5 Model adequacy
The final step after finding the parameters \(\mu\), \(\sigma_{T}^{2}\) and \(H\), is to check the adequacy of the fitted model to the data. Imagine we have a time series \(\left\{ {T_{t} } \right\}_{t = 1, \ldots ,N}\). The residuals of our fGn model are obtained from inverting Eq. (46) and calculating the vector
If the model provides a good description of the data, the elements of the residual vector \({\mathbf{e}}_{N} = \left[ {e_{1} , \ldots ,e_{N} } \right]^{T}\) should be white noise, i.e. they should be \({\text{NID}}\left( {0,1} \right)\) with autocorrelation function \(\left\langle e_{i} e_{j}\right\rangle = \delta_{ij}\). Many statistical tests for whiteness of \(\left\{ {e_{i} } \right\}\) could be performed, the more descriptive one being based on the examination of the graph of the residual autocorrelation function (RACF). The RACF at lag \(l\) is calculated as:
Asymptotically, \(r_{l} \left( {{\mathbf{e}}_{N} } \right)\sim {\text{NID}}\left( {0,1/N} \right)\) for any lag \(l \ge 1\) and \(r_{0} \left( {{\mathbf{e}}_{N} } \right) = 1\). In the graph of \(r_{l} \left( {{\mathbf{e}}_{N} } \right)\) vs. \(l\), there should not be any point significantly far outside the 95% confidence interval given by the horizontal lines \(\pm 1.96/\sqrt N\), and the number of points outside this range, should represent around 5% of the total number of points. As additional tests, we could verify that the estimates of the fluctuation exponent of \(\left\{ {e_{i} } \right\}\), using the previous graphical methods, are \(\hat{H}_{s} \approx \hat{H}_{h} \approx - 0.5\), which is the value for white noise as a particular case of fGn. The less important Gaussianity assumption could also be verified by visualizing the empirical probability distribution against a normal distribution and checking for the presence of extremes.
Appendix 2: Checking the fGn model fit to global temperature data
In Table 5 we show the values of the parameters obtained for the ten datasets and the corresponding mean series for the globe and for land:
As we can see in Table 5, there is relatively good agreement between the more robust estimates of the fluctuation exponent, \(\hat{H}_{l}\) and \(\hat{H}_{q}\) (see Appendix 1 for the notation), with the small bias of \(\hat{H}_{q}\) towards smaller values (we used a memory \(p = 20\) months for estimating \(\hat{H}_{q}\)). The estimates \(\hat{H}_{h}\) and \(\hat{H}_{s}\), obtained using the general methods, also roughly agree with the MLE and QMLE considering their relatively wide one-standard deviation confidence interval (given in parentheses in Table 5). Notice the difference between the parameter \(\hat{\sigma }_{T}\) and the amplitude of each series, \(SD_{T}\). The former is an unbiased estimate of the standard deviation for the ensemble process using maximum likelihood, while the latter is a biased estimate, where the bias is because of the limited time series and autocorrelated samples (see Ergodicity in Appendix 1). We also include the values of \(SD_{T} /\sqrt {1 - N^{2H} }\) for confirmation of Eq. (25) (\(N = 1656\) months). The last two columns show the climate sensitivity, \(\lambda_{{2 \times {\text{CO}}_{2} {\text{eq}}}}\), and the parameter \(T_{0}\) (Eq. (27)) used to remove the anthropogenic trend in each global series. The value \(T_{0}\) was chosen to obtain \(\bar{T}_{\text{nat}} = 0\) for each dataset, but this condition does not imply that \(\hat{\mu } = 0\) in Eq. (49), as this last one is an estimate for the ensemble mean. Nevertheless, the values obtained for \(\hat{\mu }\) were too small compared to \(\hat{\sigma }_{T}\) and they were not included in Table 5.
With the parameters shown in Table 5 for global temperature series, we can check the fit of the model to the data as described at the end of Appendix 1. As an example, in Fig. 22 we show the natural variability component for the Mean-G dataset, together with its corresponding series of residual innovations, \(\left\{ {e_{i} } \right\}\), obtained using Eq. (67). The first series should be Gaussian with standard deviations \(SD_{T}\) while the residuals should be white noises, i.e. they should be \({\text{NID}}\left( {0,1} \right)\) with autocorrelation function \(\left\langle e_{i} e_{j}\right\rangle = \delta_{ij}\). To verify the whiteness of the innovations, we should check that the residual autocorrelation function (RACF, (Eq. (68)) satisfies \(r_{l} \left( {{\mathbf{e}}_{N} } \right)\sim {\text{NID}}\left( {0,1/N} \right)\) for any lag \(l \ge 1\) (for \(l = 0\), \(r_{0} \left( {{\mathbf{e}}_{N} } \right) = 1\)).
The graph of the RACF for the innovations of the Mean-G dataset is shown in Fig. 23 for \(0 \le l \le N/4\), where \(N = 1656\) is the total number of points. The inset was obtained by dropping the point for zero lag and zooming in the y-axis. The theoretical 95% confidence interval, given by the values \(\pm 1.96/\sqrt N\), is shown in dashed lines. From a direct inspection, we can see that there are not too many points that fall outside the band considered and the extreme values are not too far from these thresholds.
With the purpose of checking the Gaussianity hypothesis of the series represented in Figs. 22 and 23, a detailed statistical analysis was performed. Extremes in temperature natural variability are an important issue for the prediction of catastrophic events. Its presence would show as large tails in the distributions of temperature anomalies and their corresponding innovations. If this were the case, the model could be fixed by assuming white noise with a different distribution for the innovations (e.g. a Levy distribution or one from a multifractal process). On the other hand, deviations from Gaussianity in the RACF distributions would imply a different correlation structure and would automatically invalidate the applicability of the fGn model.
As an example, in Fig. 24, we show, from top to bottom, the results of this analysis for the natural variability component of the Mean-G dataset, for its corresponding series of residual innovations and for the RACF. In the left, there is a visual comparison of the empirical cumulative distribution functions, CDF, (blue) to that of the respective fitted Gaussian distributions (red) and in the right the more enlightening probability graphs where the empirical probabilities obtained from the graphs in the left are plotted against the theoretical probabilities (blue curve). The reference line shown in red corresponds to a perfect fit. The Kolmogorov–Smirnov (K–S) test can be used to create a measure that quantifies the behavior in probability graphs. The K–S test statistic is equivalent to the maximum vertical distance between a point in the plot and the reference line. The closer the points are to the reference line, the more probable is the data satisfies the fitted theoretical distribution.
In Table 6 we summarize the standard deviations of the normal distributions obtained for the series of anomalies (\(SD_{T}\)), the series of residual innovations (\(SD_{\text{innov}}\)) and the RACF (\(SD_{\text{RACF}}\)) for each dataset. The mean values of the distributions were very small compared to the respective standard deviations and they were omitted. The K–S test statistics with the corresponding p-values are also shown. More powerful statistical tests for normality could be performed, like the Shapiro–Wilk or the Anderson–Darling tests. However, these other tests have their own disadvantages, and, for the purpose of this work, the conclusions obtained from the K–S test to check the Gaussianity hypothesis of the original anomalies and the adequacy of the fGn process fit, are good enough.
The values of \(SD_{T}\) are the same shown previously in Table 5. As expected from the theory, \(SD_{\text{innov}} = 1\) for all dataset and the values obtained for \(SD_{\text{RACF}}\) are close to the theoretical value \(1/\sqrt N = 0.025\) (\(N = 1656\)). With the exceptions of the residual innovations of NOAA and HAD4 for the global datasets, the p-values are above 0.05, so there is not enough evidence to reject normality at that level. Moreover, the p-values obtained are, in general, larger than those obtained for series of the same length based on pseudorandom number generators [for a numerical experiment using 10000 samples, the p-values were uniformly distributed in the range (0–1)]. For the land surface datasets, the p-values for the temperature anomalies and the innovations are low and a different distribution for the white noise innovations could be proposed.
As we mentioned before, the normality of the innovations is less important to confirm the adequacy of the model than its whiteness, which is confirmed by the Gaussianity of the RACF in all cases (see the large p-values in the last column of Table 6). A main deviation from normal behavior is the existence of extremes in the original data. This “fat-tailed” property of the probability distributions was evidenced in Lovejoy (2014) in a paper of statistical hypothesis testing of anthropogenic warming. In the present work, it does not have major implications or compromise the applicability of the model to the global data.
Appendix 3: Forecast and validation for all datasets
Some results of the hindcast validation are summarized in Table 7 for the twelve datasets, including the mean series for the global and the land surface. Only the error, \({\text{RMSE}}_{\text{nat}}\), and the \({\text{ACC}}_{\text{nat}}\), for the natural variability component were presented for horizons \(k =\) 1, 3, 6 and 12 months. The values \({\text{MSSS}}_{\text{nat}}\) and \({\text{MSSS}}_{\text{raw}}\) can be obtained from Eq. (34) taking \({\text{MSE}} = {\text{RMSE}}^{2}\) and the respective \({\text{MSE}}_{\text{ref}} = SD_{T}^{2}\) or \({\text{MSE}}_{\text{ref}} = SD_{\text{raw}}^{2}\). Also, we can use the values of \({\text{ACC}}_{\text{nat}}\) to obtain very good approximations of \({\text{MSSS}}_{\text{nat}}\) for these horizons thanks to the relationship \({\text{MSSS}}_{\text{nat}} \approx {\text{ACC}}_{\text{nat}}^{2}\) (Eq. (38)). Only the spurious values of \({\text{ACC}}_{\text{raw}}\) cannot be obtained from this table, but it is worth mentioning that, even for \(k = 12\) months, they are higher than 0.75 for all datasets. Notice the large difference between the values of \(SD_{T}\) and \(SD_{\text{raw}}\), for the detrended and the raw anomalies respectively, due to the presence of the anthropogenic trend. The values of \(\hat{\sigma }_{T}\), were included for reference.
Rights and permissions
About this article
Cite this article
Del Rio Amador, L., Lovejoy, S. Predicting the global temperature with the Stochastic Seasonal to Interannual Prediction System (StocSIPS). Clim Dyn 53, 4373–4411 (2019). https://doi.org/10.1007/s00382-019-04791-4
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00382-019-04791-4