Abstract
Using 5-day averaged precipitation from all initial condition realizations of 33 CMIP5 models for the Historical and RCP8.5 scenarios, we performed an assessment of summer precipitation in terms of amount, onset date, withdrawal date, and length of season using probability distributions of interannual anomalies. Climate change projections were generated using all models, one model per modelling group to account for overconfidence, and sub-selecting models on annual cycle fidelity. Compared to using all models, sub-selecting on annual cycle fidelity has a large impact on the climate change perturbation of the fractional change in precipitation, with differences between the two projections of up to ± 50%, especially in the tropics and subtropics. Sensitivity testing indicates that the Gaussian t-test and the non-parametric Mann–Whitney U-test (the latter using Monte Carlo sampling) yield consistent results for assessing where the climate change perturbation is significant at the 1% level, even in cases where skewness and excess kurtosis indicate non-Gaussian behavior. Similarly, in terms of climate change induced perturbations to below-normal, normal, and above-normal categorical probabilities, the Gaussian results are typically consistent with the non-parametric estimates. These sensitivity results promote the use of Gaussian statistics to present global maps of the lower-bound and upper-bound of the climate change response, given that the non-parametric calculation of confidence intervals would otherwise not be tractable in a desktop computing environment due to its CPU intensive requirement.
This is a preview of subscription content, access via your institution.













References
Allen JT, Tippett MK, Sobel AH (2015) Influence of the El Niño/Southern oscillation on tornado and hail frequency in the United States. Nat Geosci 8:278–283. https://doi.org/10.1038/NGEO2385
Angélil O, Stone DA, Tadross M, Tummon F, Wehner M, Knutti R (2014) Attribution of extreme weather to anthropogenic greenhouse emissions: sensitivity to spatial and temporal scales. Geophys Res Lett 41:2150–2155. https://doi.org/10.1002/2014GL059234
Annamalai H, Hamilton K, Sperber KR (2007) The South Asian summer monsoon and its relationship with ENSO in the IPCC AR4 simulations. J Clim 20:1071–1092. https://doi.org/10.1175/JCLI4035.1
Basistha A, Arya DS, Goel NK (2009) Analysis of historical changes in rainfall in the Indian Himalayas. Int J Climatol 29:555–572. https://doi.org/10.1002/joc.1706
Biasutti M (2013) Forced Sahel rainfall trends in the CMIP5 archive. J Geophys Res Atmos 118:1613–1623. https://doi.org/10.1002/jgrd.50206
Bombardi RJ, Carvalho LMV (2009) IPCC global coupled model simulations of the South America monsoon system. Clim Dyn 33:893–916. https://doi.org/10.1007/s00382-008-0488-1
Brown JR, Colman RA, Moise AF, Smith IN (2013) The western Pacific monsoon in CMIP5 models: model evaluation and projections. J Geophys Res Atmos 118:12,458–12,475. https://doi.org/10.1002/2013jd020290
Challinor AJ, Watson J, Lobell DB, Howden SM, Smith DR, Chhetri N (2014) A meta-analysis of crop yied under climate change and adaptation. Nat Clim Change 4:287–291. https://doi.org/10.1038/NCLIMATE2153
Chou C, Neelin JD, Chen C-A, Tu J-Y (2009) Evaluating the “Rich-Get-Richer” mechanism in tropical precipitation change under global warming. J Clim 22:1982–2005. https://doi.org/10.1175/2008JCLI2471.1
Coleman RA, Moise AF, Hanson LI (2011) Tropical Australian climate and Australian monsoon as simulated by 23 CMIP3 models. J Geophys Res 116:D10116. https://doi.org/10.1029/2010JD015149
Collins M, Knutti R, Arblaster J, Dufresne J-L, Fichefet T, Friedlingstein P, Gao X, Gutowski WJ, Johns T, Krinner G, Shongwe M, Tebaldi C, Weaver AJ, Wehner M (2013) Long-term climate change: projections, commitments and irreversibility. In: Stocker TF, Qin D, Plattner G-K, Tignor M, Allen S, Boschung J, Nauels A, Xia Y, Bex V, Midgley PM (eds) Climate change 2013: the physical science basis. Contribution of working group I to the fifth assessment report of the intergovernmental panel on climate change. Cambridge University Press, Cambridge
Cook BI, Seager R (2013) The response of the North American monsoon to increased greenhouse gas forcing. J Geophys Res 118:1690–1699. https://doi.org/10.1002/jgrd.50111
Craig MH, Snow RW, le Sueur D (1999) A climate-based distribution model of Malaria transmission in Sub-Saharan Africa. Parasitol Today 15:105–111
Divine GW, Norton HJ, Barón AE, Juarez-Colunga E (2018) The Wilcoxon–Mann–Whitney procedure fails as a test of medians. Am Stat 72:278–286. https://doi.org/10.1080/00031305.2017.1305291
Ersek V et al (2012) Holocene winter climate variability in mid-latitude western North America. Nat Comm 3:1219. https://doi.org/10.1038/ncomms2222
Fischer EM, Sedlacek J, Hawkins E, Knuti R (2014) Models agree on forced response pattern of precipitation and temperature extremes. Geophys Res Lett 41:8554–8562. https://doi.org/10.1002/2014GL062018
Flato G, Marotzke J, Abiodun B, Braconnot P, Chou SC, Collins W, Cox P, Driouech F, Emori S, Eyring V, Forest C, Gleckler P, Guilyardi E, Jakob C, Kattsov V, Reason C, Rummukainen M (2013) Evaluation of climate models. In: Stocker TF, Qin D, Plattner G-K, Tignor M, Allen SK, Boschung J, Nauels A, Xia Y, Bex V, Midgley PM (eds) Climate Change 2013: the physical science basis. Contribution of Working Group I to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change. Cambridge University Press, Cambridge
Furrer R, Knutti R, Sain SR, Nychka DW, Meehl GA (2007) Spatial patterns of probabilistic temperature change projections from a multivariate Bayesian analysis. Geophys Res Lett 34:L06711. https://doi.org/10.1029/2006/GL027754
Gleckler PJ, Taylor KE, Doutriaux C (2008) Performance metrics for climate models. J Geophys Res 113:D06104. https://doi.org/10.1029/2007JD008972
Greene AM, Goddard L, Lall U (2006) Probabilistic multimodel regional temperature change projections. J Clim 19:4326–4343
Held IM, Soden BJ (2006) Robust responses of the hydrological cycle to global warming. J Clim 19:5686–5699
Huffman GJ, Adler RF, Arkin P, Chang A, Ferraro R, Gruber A, Janowiak J, McNab A, Rudolf B, Schneider UJ (1997) The global precipitation climatology project (GPCP) combined precipitation dataset. Bull Am Meteor Soc 78:5–20
Jacob D et al (2014) EURO-CORDEX: new high-resolution climate change projections for European impact research. Reg Environ Change 14:563–578. https://doi.org/10.1007/s10113-013-0499-2
Jones C, Carvahlo LMV (2013) Climate change in the South American monsoon system: present climate and CMIP5 projections. J Clim. https://doi.org/10.1175/JCLI-D-12-00412.1
Kharin VV, Zwiers F, Zhang X, Wehner M (2013) Changes in temperature and precipitation extremes in the CMIP5 ensemble. Clim Change 119:345–357. https://doi.org/10.1007/s10584-013-0705-8
Kitoh A (2017) The Asian monsoon and its future change in climate models: a review. J Meteorol Soc Japan 95:7–33. https://doi.org/10.2151/jmsj.2017-002
Kitoh A, Endo H, Krishna Kumar K, Cavalcanti IFA, Goswami P, Zhou T (2013) Monsoons in a changing world: a regional perspective in a global context. J Geophys Res Atmos 118:3053–3065. https://doi.org/10.1002/jgrd.50258
Knutti R, Furrer R, Tebaldi C, Cermak J, Meehl GA (2010) Challenges in combining projections from multiple models. J Clim 23:2739–2758. https://doi.org/10.1175/2009/JCLI3361.1
Leith C (1973) The standard error of time-average estimates of climatic mean. J App Meteorol 12:1066–1069
Lopez A, Tebaldi C, New M, Stainforth D, Allen M, Kettleborough J (2006) Two approaches to quantifying uncertainty in global temperature changes. J Clim 19:4785–4796
Maloney E et al (2014) North American climate in CMIP5 experiments: part III: assessment of twenty-first-century projections. J Clim 27:2230–2270. https://doi.org/10.1175/JCLI-D-13-00273.s1
Mann HB, Whitney DR (1947) On a test of whether one of two random variables is stochastically larger than the other. Ann Mathemat Stat 18:50–60
Myhre G, Shindell D, Bréon F-M, Collins W, Fuglestvedt J, Huang J, Koch D, Lamarque J-F, Lee D, Mendoza B, Nakajima T, Robock A, Stephens G, Takemura T, Zhang H (2013) Anthropogenic and natural radiative forcing. In: Stocker TF, Qin D, Plattner G-K, Tignor M, Allen SK, Boschung J, Nauels A, Xia Y, Bex V, Midgley PM (eds) Climate Change 2013: the physical science basis. Contribution of Working Group I to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change. Cambridge University Press, Cambridge
Pascale S, Boos WR, Bordoni S, Delworth TL, Kapnick SB, Murakami H, Vecchi GA, Zhang W (2017) Weakening of the North American monsoon with global warming. Nat Clim Change 7:806–813. https://doi.org/10.1038/nclimate3412
Porter JR, Xie L, Challinor AJ, Cochrane K, Howden SM, Iqbal MM, Lobell DM, Travasso MI (2014) Food security and food production systems. In: Field CB, Barros VR, Dokken DJ, Mach KJ, Mastrandrea MD, Bilir TE, Chatterjee M, Ebi KL, Estrada YO, Genova RC, Girma B, Kissel ES, Levy AN, MacCracken S, Mastrandrea PR, White LL (eds) Climate Change 2014: impacts, adaptation, and vulnerability. Part A: global and sectoral aspects. Contribution of Working Group II to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change. Cambridge University Press, Cambridge, pp 485–533
Räisänen J, Palmer TN (2001) A probability and decision-model analysis of a multimodel ensemble of climate change simulations. J Clim 14:3212–3226
Reisinger A, Kitching RL, Chiew F, Hughes L, Newton PCD, Schuster SS, Tait AA, Whetton P (2014) Australasia. In: Barros VR, Field CB, Dokken DJ, Mastrandrea MD, Mach KJ, Bilir TE, Chatterjee M, Ebi KL, Estrada YO, Genova RC, Girma B, Kissel ES, Levy AN, MacCracken S, Mastrandrea PR, White LL (eds) Climate change 2014: impacts, adaptation, and vulnerability. Part B: regional aspects. Contribution of Working Group II to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change. Cambridge University Press, Cambridge, pp 1371–1438
Rodwell MJ, Hoskins BJ (1996) Subtropical anticyclones and summer monsoons. J Clim 14:3192–3211
Sanderson BM, Knutti R, Caldwell P (2015) Addressing interdependency in a multimodel ensemble by interpolation of model properties. J Clim 28:5150–5170. https://doi.org/10.1175/JCLI-D-14-00361.1
Sardeshmukh P, Compo GP, Penland C (2015) Need for caution in interpreting extreme weather statistics. J Clim 28:9166–9187. https://doi.org/10.1175/JCLI-D-15-0020.1
Sawilowsky SS, Blair RC (1992) A more realistic look at the robustness and type II error properties of the t-test to departures from population normality. Psychol Bull 111:352–360
Singh R, AchutaRao K (2018) Quantifying uncertainty in twenty-first century climate change over India. Clim Dyn. https://doi.org/10.1007/s00382-018-4361-6
Siraj AS, Santos-Vega M, Bouma MJ, Yadeta D, Carrascal DR, Pascual M (2014) Altitudinal changes in malaria incidence in highlands of Ethiopia and Colombia. Science 343(6175):1154–1158. https://doi.org/10.1126/science.1244325
Smith KR, Woodward A, Campbell-Lendrum D, Chadee DD, Honda Y, Liu Q, Olwoch JM, Revich B, Sauerborn R (2014) Human health: impacts, adaptation, and co-benefits. In: Field CB, Barros VR, Dokken DJ, Mach KJ, Mastrandrea MD, Bilir TE, Chatterjee M, Ebi KL, Estrada YO, Genova RC, Girma B, Kissel ES, Levy AN, MacCracken S, Mastrandrea PR, White LL (eds) Climate change 2014: impacts, adaptation, and vulnerability. Part A: global and sectoral aspects. Contribution of Working Group II to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change. Cambridge University Press, Cambridge, pp 709–754
Sperber KR, Annamalai H (2014) The use of fractional accumulated precipitation for the evaluation of the annual cycle of monsoons. Clim Dyn 43:3219–3244. https://doi.org/10.1007/s00382-014-2099-3
Sperber KR, Annamalai H, Kang I-S, Kitoh A, Moise A, Turner A, Wang B, Zhou T (2013) The Asian summer monsoon: an intercomparison of CMIP5 vs. CMIP3 simulations of the late 20th century. Clim Dyn 41:2711–2744. https://doi.org/10.1007/s00382-012-1607-6
Sperber KR, Cusiner E, Kitoh A, Mechoso CR, Moise A, Moufouma-Okia W, Schiro K, Turner AJ (2016) Modelling monsoons. In: Chang CP (ed) The global monsoon system: research and forecast. World scientific, Singapore
Taxak AK, Murumkar AR, Arya DS (2014) Long term spatial and temporal rainfall trends and homogeneity analysis in Wainganga basin, Central India. Weather Clim Extrem 4:50–61. https://doi.org/10.1016/j.wace.2014.04.005
Taylor KE, Stouffer RJ, Meehl GA (2012) An overview of CMIP5 and the experiment design. Bull Am Meteorol Soc 93:485–498. https://doi.org/10.1175/BAMS-D-11-00094.1
Tebaldi C, Smith RL, Nychka D, Mearns LO (2005) Quantifying uncertainty in projectioins of regional climate change: a Bayesian approach to analysis of multimodel ensembles. J Clim 18:1524–1540
Thompson MC, Doblas-Reyes FJ, Mason SJ, Hagedorn R, Conner SJ, Phindela T, Morse AP, Palmer TN (2006) Malaria early warnings based on seasonal climate forecasts from multi-model ensembles. Nature 439:576–579. https://doi.org/10.1038/nature04503
Turner AG, Annamalai H (2012) Climate change and the south Asian summer monsoon. Nat Clim Change 2:5857–5950. https://doi.org/10.1038/nclimate1495
van Oldenborgh GJ, Philip SY, Collins M (2005) El Nino on a changing climate: a multi-model study. Ocean Sci 1:81–95
Vizy EK, Cook KH, Cretat J, Neupane N (2013) Projections of a wetter Sahel in the Twenty-First century from global and regional models. J Clim 26:4664–4687. https://doi.org/10.1175/JCLI-D-12-00533.1
Wang B, Yim S-Y, Lee J-Y, Liu J, Ha K-J (2014) Future changes of Asian–Australian monsoon under RCP4.5 anthropogenic warming scenario. Clim Dyn 42:83–100. https://doi.org/10.1007/s00382-013-1769-x
Wilks DS (2006) Statistical methods in the atmospheric sciences, 2nd edn. Elsevier Publishing, Amsterdam, p 630
Wu Z, Schneider EK, Kirtman BP, Sarachik ES, Huang NE, Tucker CJ (2008) The modulated annual cycle: an alternative reference frame for climate anomalies. Clim Dyn 31:823–841. https://doi.org/10.1007/s00382-008-0437-z
Xie PP, Arkin PA (1997a) Global pentad precipitation analysis based on gauge observations, satellite estimates and model outputs. Extended Abstracts, Amer. Geophys. Union 1997 Fall Meeting, San Francisco, CA, Amer. Geophys. Union
Xie PP, Arkin PA (1997b) Global precipitation: a 17-year monthly analysis based on gauge observations, satellite estimates, and numerical model outputs. Bull Am Meteorol Soc 78:2539–2550
Xie PP, Janowiak JE, Arkin PA, Adler R, Gruber A, Ferraro R, Huffman GJ, Curtis S (2003) GPCP pentad precipitation analyses: an experimental dataset based on gauge observations and satellite estimates. J Clim 16:2197–2214
Acknowledgements
K. R. Sperber and G. Pallotta were supported by the Office of Science (BER), U.S. Department of Energy through Lawrence Livermore National Laboratory contract DE-AC52-07NA27344. H. Annamalai was supported by National Science Foundation (NSF) award 1460742. We acknowledge the World Climate Research Programme’s Working Group on Coupled Modelling, which is responsible for CMIP, and we thank the climate modeling groups (listed in Table 1 of this paper) for producing and making available their model output. For CMIP, the U.S. Department of Energy’s Program for Climate Model Diagnosis and Intercomparison provides coordinating support and led development of software infrastructure in partnership with the Global Organization for Earth System Science Portals.
Disclaimer
This document was prepared as an account of work sponsored by an agency of the United States government. Neither the United States government nor Lawrence Livermore National Security, LLC, nor any of their employees makes any warranty, expressed or implied, or assumes any legal liability or responsibility for the accuracy, completeness, or usefulness of any information, apparatus, product, or process disclosed, or represents that its use would not infringe privately owned rights. Reference herein to any specific commercial product, process, or service by trade name, trademark, manufacturer, or otherwise does not necessarily constitute or imply its endorsement, recommendation, or favoring by the United States government or Lawrence Livermore National Security, LLC. The views and opinions of authors expressed herein do not necessarily state or reflect those of the United States government or Lawrence Livermore National Security, LLC, and shall not be used for advertising or product endorsement purposes.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix: methods
Appendix: methods
1.1 A.1 Terminology
In Sperber and Annamalai (2014) we used the term “duration” to represent the time from onset through withdrawal. However, herein we prefer the term “accumulation time” to avoid confusion with other climate change studies in which “duration” is defined differently (e.g., Kitoh et al. 2013). Herein, a shorter (longer) accumulation time means that it takes less (more) time under RCP.5 compared to the Historical simulation to achieve the fractional accumulation that occurred between observed pentads 31 and 55 (Fig. 1a–c).
1.2 A.2 Annual cycle fidelity sub-selection
The development of optimal strategies for combining climate projections is still in its “infancy” (Knutti et al. 2010), with three approaches typically used in the literature. Approach 1, using all available models with equal weighting, is the dominant method of generating projections (e.g., Kitoh et al. 2013). Numerous reasons exist for using all models, with (a) this being the most straightforward method to employ, (b) recognizing the effort made by all modelling groups to produce the necessary simulations, and (c) the assumption that using all models results in the best estimate of the climatological mean state projections by extension of the fact that the multi-model mean in present-day simulations tends to produce better agreement with observed climatologies than any individual model (Gleckler et al. 2008). Approach 2 addresses model independence considerations by producing projections that use one model version per modelling group (Fischer et al. 2014) to address the issue of overconfidence (Knutti et al. 2010), though this does not address the impact of model interdependence between different modelling groups (Sanderson et al. 2015). Approach 3 is physically based, with the sub-selection based on model fidelity in representing the mode of variability in question, with a uniform subset of models being used at all gridpoints. For example, this approach has been taken with respect to ENSO (Annamalai et al. 2007; van Oldenborgh et al. 2005) and the global monsoon (Wang et al. 2014). Approach 4, used in this paper, is a variant of Approach 3, but we sub-select models based on their annual cycle fidelity at each gridpoint, since it has been demonstrated that no single CMIP5 model performs well in terms of onset, withdrawal, and accumulation time of fractional precipitation over all monsoon domains (Sperber and Annamalai 2014). Spatial heterogeneity of the annual cycle of precipitation within monsoon domains has also been found, with Bombardi and Carvalho (2009) finding that the annual cycle is better represented over central and eastern Brazil compared to the Amazon and northwest South America in 10 CMIP3 models. Annual cycle errors have also been found over the Sahel (Vizy et al. 2013) and Australia (Coleman et al. 2011) in CMIP simulations. Even though results suggest that sub-selecting on model quality only weakly relates to the projected change (Knutti et al. 2010), our conservative approach is to sub-select models that are the physically realistic in terms of the timing of summer precipitation and its interannual variability in their historical simulations, since these quantities are directly relevant to the projected precipitation characteristics we are evaluating. In this way, we exclude models that have unrealistic annual cycle phasing of summer precipitation, and unrealistic levels of interannual variability, which might otherwise compromise the projections.
To subset the models we evaluate the onset pentad and accumulation time of summer rainfall in the Historical simulations. Consideration is also given to the standard deviations of the onset and accumulation times, with the goal of not including models whose Historical simulations are excessively under-dispersive or over-dispersive compared to GPCP and CMAP. Relevant to the discussion of model quality vs. climate change perturbations for area-averaged data (Sect. 5.1), in Table 2 are the means and standard deviations (σ) of the onset pentad and accumulation time (pentads) from GPCP, CMAP, and the CMIP5 Historical simulations for India (AIR), Sahel, Northern tropical Australia (Aus), North American Monsoon (NAM), South American Monsoon (SAM), and the southern subtropical Atlantic Ocean (SAtl).
The iterative procedure for sub-selecting models is based on performance of the CMIP5 Historical simulations relative to an “Optimal Range” for onset time, accumulation time, and their variances derived from GPCP and CMAP (1979–2004) with the:
Initial “Optimal Range” from observations:
Onset time: mean ± 1σ.
Accumulation time: mean ± 1σ.
Onset variance: model variance must be within a factor of 1/2.52–2.73 of the observed variance (limits correspond to an F-test at the 1% level for 26 years of observations and 39 years of model data).
Accumulation time variance: model variance must be within a factor of 1/2.52–2.73 of the observed variance.
If a model fails any of the criteria, the optimal ranges of the onset time and accumulation time are broadened by an additional ± 0.1σ up to a
maximum range = observed mean ± 1σ ± 1 pentad.
If a model still fails any of the criteria, the maximum ranges of the onset time and accumulation time are used, and the variance range is incremented by 0.1 to a
maximum range = 1/2.79–3.
If at any iteration 6 or more models meet the sub-selection criteria, the iterative process is stopped and the climate change impact for this subset of models is calculated.
If the criteria are not met after relaxing the “Optimal Range” to their maximum ranges, climate change is not assessed at these gridpoints and the output is set to missing.
Note: GPCP has missing data at high northern latitudes and off the coast of Antarctica (shaded gray in Fig. 1d), hence no analysis is performed at these gridpoints.
1.3 A.3 Statistical analysis based on the Gaussian approach
We assess (1) changes in the mean by using a two-sided t-test at the 1% significance level, (2) determine the associated 99% confidence interval, and (3) evaluate changes in the probabilities of below-normal, normal, and above-normal tercile categories (Leith 1973) of the PDFs of interannual precipitation anomalies in the RCP8.5 simulations relative to those from the Historical simulations. For the case of having one initial condition realization per model we could calculate the perturbation to the tercile categories under RCP8.5 relative to the Historical simulations directly from the anomalies without assuming an underlying statistical model. However, because we have differing numbers of realizations from the individual models, using the full suite of anomalies in such an empirical approach would bias the probability estimates of the climate change signal toward models with more than one realization.
To ensure equal weighting for each model, for statistical purposes we assume the anomalies for each model to be Gaussian distributed. For each model realization we calculate Historical anomalies for each year with respect to that realization mean. For the RCP8.5 realizations performed using that model, the anomalies are calculated with respect to the ensemble mean of its available Historical realizations. For the evaluation of fractional changes in precipitation, we also divide the anomalies by the Historical ensemble mean. We then calculate the mean, variance, and the PDF of the Historical anomalies (mean = 0 by definition) and of the RCP8.5 anomalies. Then, over all models, we calculate the ensemble averages of the means, variances, and the PDFs so that each model is equally weighted. Support for this approach is given in Brown et al. (2013), who also calculated the PDFs for each model and averaged them together to create the multi-model mean PDF. We then use the resulting ensemble mean and variance to perform the t-test to assess if the mean of RCP8.5 is statistically different from the mean of the Historical simulations (see Sect. 5). Under the assumption that the anomalies are Gaussian distributed, using Simpson’s Rule we analytically calculate the tercile categorical probabilities in the RCP8.5 simulations relative to those defined from the Historical simulations. These probabilistic changes in RCP8.5 are assessed using cutoff thresholds for three equally probable categories determined from the Historical anomalies (Leith 1973). Using the Historical anomalies, the below-normal category is defined as − 6 standard deviations to − 0.43073 standard deviations, the normal category spans ± 0.43073 standard deviations, and the above-normal category is defined for 0.43073 standard deviations to 6 standard deviations. The pooled variance is used for calculating the categorical changes. Irrespective of the actual shapes of the multi-model PDFs seen in the accompanying histograms, all statistics are calculated using the Gaussian assumption. Hereafter, we refer to this procedure as the “Gaussian approach”. In the Gaussian approach, for each realization we calculate the effective degrees of freedom by taking serial correlation into account, since serial correlation inflates the standard deviation when the lag 1 autocorrelation is greater than zero (Wilks 2006). The total degrees of freedom equals the sum of the effective degrees of freedom using all realizations.
1.4 A.4 Non-parametric analysis using the Monte Carlo sampling approach
While the calculation of the afore-mentioned Gaussian approach is computationally very fast, it is also instructive to evaluate the climate change signals using a non-parametric approach, since Sardeshmukh et al. (2015) reported that seasonal mean GPCP precipitation exhibits strong skewness, especially in regions of descent. Though Sawilowsky and Blair (1992) demonstrate that the two-tailed t-test yields conservative results for large sample sizes (~ 60 or more samples) when performed at the 1% significance level, even for PDFs that are highly skewed, non-Gaussianity of the PDFs will impact the tercile probabilities for the below-normal, normal, and above-normal categories. Thus, we also report results using a Monte Carlo sampling using non-parametric statistics. In the cases in which multiple realizations are available for a given model (Table 1) we randomly select one Historical realization and one RCP8.5 realization. The random selection is performed separately for the Historical and for the RCP8.5 realizations, and at each gridpoint for the spatial patterns.
Based on the anomalies from the selected models (33 models, 18 models, or the annual cycle fidelity subset), for each Monte Carlo sample the two-sided Mann–Whitney U-test is performed (Mann and Whitney 1947), with the climate change signal being the median of the RCP anomalies minus the median of the Historical anomalies and with the associated P-value being retained as a measure of the statistical significance of the difference. The Mann–Whitney U-test is a “location” test, a non-parametric alternative to the t-test, with the null hypothesis being that the two batches of data (Historical and RCP8.5) are drawn from the same PDF. It operates by comparing the joint ranking of anomalies of the Historical versus RCP8.5 simulations. When anomalies in both sets of data are equal, the average rank is used to calculate a corrected standard deviation (https://en.wikipedia.org/wiki/Mann–Whitney_U_test). Working in rank-space guards against significant ensemble average changes that might arise due to anomalies that are outliers (Wilks 2006). The benefit of this test is that it is known to have higher power compared to the t-test in the presence of skewed distributions. It has been traditionally used to detect shifts in the distribution of the variables under different conditions (e.g., Basistha et al. 2009; Ersek et al. 2012; Jacob et al. 2014; Taxak et al. 2014). Although a very recent paper has raised warnings about the use of the Mann–Whitney procedure as a test of medians (Divine et al. 2018), the test still remains a good non-parametric alternative to the t-test, given that a pre-check is performed on the equality of the shape (e.g., variance) of the distributions to be compared. Before applying the Mann–Whitney U-test herein, we performed an F-test on the global gridded. For example, for the fractional change in precipitation we find that over most of the globe we cannot reject the null hypothesis that the Historical and RCP variances are equal at the 1% significance level for ≥ 70% of the models (not shown). The main exceptions are the equatorial central/eastern Pacific, poleward of about 80°N, and the interior of Antarctica over the eastern hemisphere where the null hypothesis cannot be rejected for ≤ 40% of the models. As part of the Monte Carlo testing we also perform a t-test in which we calculate the effective degrees of freedom by taking serial correlation into account. The t-test results can be compared to the non-parametric Mann–Whitney U-test results, and to the Gaussian approach obtained using the Section A.3 methodology. We also empirically calculate the tercile probabilities of the RCP8.5 anomalies relative to the Historical thresholds.
Using all Monte Carlo samples we separately average the results from the Mann–Whitney U-test, the t-test, and the empirically calculated RCP8.5 tercile probabilities. In the case of 33 models, the number of Historical (RCP8.5) realizations results in 4.53 × 109 (2.04 × 106) independent combinations. However, the Monte Carlo sampling is only repeated 2000 times, which balances CPU time (~ 36 h for 1 variable for the global gridpoint data on a top-of-the-line 2018 MacBook Pro using 1 processor) vs. consistency of the results to the number of Monte Carlo samples. In the sensitivity test in which a subset of models is retained based on annual cycle fidelity, the limited number of models at some gridpoints limits the number of possible Monte Carlo samples to less than 2000. Sensitivity of the results to the number of Monte Carlo samples is explored in Sect. 5.1. Additionally, the Monte Carlo analysis has been repeated using a different “seed” to initialize the random number generator, thus allowing us to assess the consistency of our results to the selection of independent combinations of realizations.
Note: While a confidence interval for the change in the median can be estimated non-parametrically, the present desktop computing environment precludes this calculation for the global gridpoint analysis. We do present non-parametric confidence intervals for select area-averaged time series, but this is limited to at most 2000 Monte Carlo samples using at most 14 models (39 [40] yearly anomalies for each Historical [RCP8.5] realization from each model) from the annual cycle fidelity subset. In this instance, for one area-averaged dataset, the CPU time is ~ 50 h. For the non-parametric confidence interval calculation using more data becomes even more problematic, since the CPU time scales as the square of the number of possible differences between the RCP and Historical anomalies due to the sorting needed to rank the differences.
Rights and permissions
About this article
Cite this article
Sperber, K.R., Annamalai, H. & Pallotta, G. CMIP5: a Monte Carlo assessment of changes in summertime precipitation characteristics under RCP8.5-sensitivity to annual cycle fidelity, overconfidence, and gaussianity. Clim Dyn 54, 1699–1727 (2020). https://doi.org/10.1007/s00382-019-05082-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00382-019-05082-8
Keywords
- CMIP5
- Anthropogenic climate change
- Precipitation
- Annual cycle skill
- Monte Carlo sampling