1 Introduction

In recent years, society has faced a steep rise in economic and insured losses from weather and climate related hazards, largely due to significant increase in exposure (e.g. Kunreuther and Michel-Kerjan 2009). Projections of a continuing trend towards more intense systems (see Knutson et al. 2010; Holland and Bruyère 2013 for the case of tropical cyclones) point to a further increase in societal vulnerability. More accurate information on high-impact events, is thus a critical need of society. This requires assessments of the statistics of high-impact events and the associated uncertainty with regional clarity together with potential changes under climate variability and change.

Meeting these demands requires a combination of dynamical and statistical components. The traditional dynamical approach combines the capacity of regional high resolution to simulate weather events with the capacity of global coarse resolution to simulate climate by embedding high resolution within the global mesh over regions of interest (e.g. Laprise et al. 2008; Knutson et al. 2007). Increases in computational capacity have enabled such simulations in unprecedented detail (e.g. Bender et al. 2010). Regional modeling studies are subject to uncertainty and careful consideration is required of the balance between model complexity to resolve the relevant physical processes, ensemble size to sample uncertainty in the driving data, and simulation length to capture the full distribution. These competing demands for computational resources often results in a truncation of the full distribution of high-impact events and this has led to exploration of a variety of solutions including the use of empirical relationships between weather systems and the large-scale environment (e.g. Camargo et al. 2007), and assessment of errors in frequency distributions of weather events from dynamical models (e.g. Katz 2010).

Meeting the societal demand of assessing high-impact events with regional clarity remains extremely challenging. This paper presents an overview of recent lessons learnt in modeling high-impact events on regional scales using the case of tropical cyclones. Tropical cyclones represent a hard test case for simulating high-impact weather owing to their rarity, regional variability and uncertain future regional changes (see for example large variations between the modeling studies of Knutson et al. 2007, 2010 and Bengtsson et al. 2007). The merits and limitations of the dynamical modeling approach are discussed and motivate, by way of illustrative examples, the need for combined dynamical-statistical approaches in order to provide credible and useful information.

The next section describes simulation experiments with the Nested Regional Climate Model (NRCM), a dynamical downscaling tool based on the Weather Research and Forecasting (WRF) model (Skamarock et al. 2008) designed specifically to contribute to assessments of high-impact events. Atmosphere-only NRCM simulations of current and future North Atlantic tropical cyclone activity are presented to illustrate current limitations and sensitivities of the dynamical model approach. The value of complementary statistical approaches is illustrated in Section 3. Finally, key findings are summarized in Section 4.

2 Dynamical assessments

Here multi-year NRCM simulations of high-impact weather are assessed using limited area domains to identify the impact of climate bias, domain size and resolution; and to explore and document critical sensitivities of this dynamical modeling approach.

Global climate data are provided by Community Climate System Model (CCSM, Collins et al. 2006) version 3 using the A2 scenario (IPCC SRES SPM 2000), from the Coupled Model Intercomparison Project 3 (CMIP3, Meehl et al. 2007). The CCSM is a full Earth system model, including atmosphere, ocean, cryosphere, biosphere, and land surface. These data are used to drive a NRCM 36 km domain using one-way nesting, which in turn is used to drive a 12 km domain (Fig. 1) to explore the additional information, if any, provided by the higher resolution. One-way nesting is chosen to avoid the unknown implications of using two-way nesting and for the practical reason of running each domain separately and checking the simulation prior to further downscaling. A description of the NRCM and physics used is provided in the supplemental online material (SOM) section S1.

Fig. 1
figure 1

NRCM model domains at 36 km grid spacing (large black box) and 12 km grid spacing (small black box). Model terrain height (shaded) is shown at the different model resolutions and extends beyond the 36 km domain to indicate the resolution of the driving CCSM. Adapted from Done et al (2011). Copyright 2011 OTC. Reproduced with permission of the copyright owner

The simulations cover three periods: a decade of ‘current’ climate conditions (1995–2005) referred to hereafter as ‘base climate’, and two future decades of 2020–2030 and 2045–2055. The time periods are nominal since the driving CCSM model was initialized in 1950 with no additional assimilated data. Thus, for example, model interannual and multidecadal variations are not expected to match those in the real world, though the historical anthropogenic trend associated with greenhouse gases should be captured.

2.1 Domain size, location and horizontal resolution

Domain size and horizontal resolution are key factors for regional climate simulation (e.g. Vannitsem and Chome 2005) and tropical cyclone simulation (e.g. Kumar et al. 2011). High resolution is required to adequately resolve the intensity of strong tropical cyclones (Bender et al 2010), particularly if they are small. Resolution also affects cyclone structure (Fig. 2). At 36 km the simulated cyclone has a simple circular structure whereas at 12 km sub-system-scale structure emerges in the form of eye-wall asymmetries and spiral rain bands. As the grid-spacing is reduced further the shift from parameterized towards explicit convection can additionally impact other cyclone characteristics including; genesis mechanism (Kieu and Zhang 2008), eye-wall replacement cycles and rapid intensification (Davis et al. 2008), and upscale impacts through vertical redistributions of heat and momentum (as discussed in Leung et al. 2006). Finally, cyclone frequency is also found to be sensitive to resolution (SOM text, section S3). Whether this is due to changes in small- or large-scale processes is unknown. Caron et al. (2010) and Caron and Jones (2011) found high resolution was necessary for the intensity of easterly waves and even for accurate representation of the large-scale environment over the eastern Atlantic.

Fig. 2
figure 2

Snapshots of example NRCM simulated tropical cyclones in the Gulf of Mexico generated on (left) the 36 km grid and (right) the 12 km grid, shown in model derived radar reflectivity (dBz)

Large nested domains have been shown to improve the simulation of all scales within the domain interior above those in the driving global data (Jones et al. 1995; Laprise et al. 2008) and reduce spatial spin-up issues by moving the inflow boundary far from the region of interest (Leduc and Laprise 2009). Critically, the domain needs to be sufficiently large to capture regional physical processes (Giorgi and Mearns 1999) not only to assess high-impact weather in current climate but also to capture the correct climate response. A domain that is smaller than the main external modes of variability is closely coupled to the driving model (somewhat similar to nudging), and where small scales are important to these modes, the domain size needs to be large to enable this upscale interaction to occur. For example, Caron and Jones (2011) constructed a domain to capture relationships between Atlantic SSTs, Sahelian rainfall and tropical cyclogenesis.

Guided by these findings, available resources for this study are directed to domain size and resolution at the expense of run length and ensemble size so a single simulation is run on each domain. The 36 and 12 km domains are far larger than North America, our target region, to ensure the majority of atmospheric processes that impact the region are handled by the higher resolution model rather than the coarser climate model. As a specific example, African easterly waves are not well captured by the CCSM simulation, necessitating the inclusion of the African wave source and development region within the 36 km domain (see SOM section S3 for discussion on the representation of easterly waves).

2.2 Climate bias

Global climate models commonly contain climate bias and despite the large domain, the NRCM-generated climate driven directly by CCSM data contains substantial biases. Anomalously strong large-scale flow at upper-levels over the tropical North Atlantic produces strong vertical wind shear, defined as the difference in winds between 200- and 850-hPa (Fig. 3a), thereby preventing tropical cyclogenesis. Sensitivity studies (not shown) reveal that this bias is transferred to the NRCM from CCSM, in part due to dynamical propagation from the east and west boundaries but also due to a warm eastern Pacific Ocean SST bias (Large and Danabasoglu 2006). This permanent El Niño-like condition contributed to the high vertical wind shear in the NRCM over the tropical Atlantic through a modified Walker Circulation (e.g. Gray 1984). One approach to correcting bias is to statistically correct the NRCM model output (Dosio and Paruolo 2011) but here this post-processing approach is not suitable since the biased climate did not generate any tropical cyclones, thereby necessitating bias correction of the driving CCSM data prior to NRCM simulation.

Fig. 3
figure 3

Three-month average vertical wind shear (200–850 hPa, ms−1) for the period August–October 1996 for a NRCM driven by raw CCSM data; b NRCM driven by revised CCSM data; c raw CCSM data, and d NNRP data. Adapted from Holland et al (2010). Copyright 2010 OTC. Reproduced with permission of the copyright owner

A popular approach to assessing regional climate change that implicitly removes bias is the pseudo-global-warming approach (Schär et al. 1996; Rasmussen et al. 2011). In this approach, reanalysis data are used for current climate and future climate is constructed by adding a perturbation, intended to represent the mean climate change, to the reanalysis data. This approach assumes no change in variability at the domain boundaries and for small domains will constrain the frequency of weather events to current climate only. An alternative bias correction technique is used here that allows synoptic and climate variability to change in the future. Six-hourly CCSM data for the entire simulation (1950–2100) are broken down into an average annual cycle plus a perturbation term:

$$ CCSM=\overline{ CCSM}+ CCSM\prime, $$
(1)

where CCSM′ varies in time throughout the entire 150-year CCSM simulation period and includes both high-frequency variability and climate trends. The average annual cycle is defined over a 20-year base period (to smooth out any influence of El Niño) from 1975 to 1994. Twenty years may not be sufficient to smooth out influence of multi-decadal variability but is chosen to avoid inclusion of any climate trends. Similarly, 6-hourly NCEP/NCAR Reanalysis Project (NNRP, Kalnay et al. 1996) data for 1975–1994 are broken down into an average annual cycle plus a perturbation term:

$$ NNRP=\overline{ NNRP}+ NNRP\prime, $$
(2)

The revised climate data, CCSM R , are then constructed by replacing \( \overline{ CCSM} \) in Eq. 1 with \( \overline{ NNRP} \) in Eq. 2:

$$ CCS{M}_R=\overline{ NNRP}+ CCSM\prime, $$
(3)

CCSM R therefore combines a base, seasonally varying climate provided by reanalysis data with day-to-day weather, climate variability (e.g. synoptic weather systems, ENSO) and climate change provided by CCSM. This approach is similar to the method used by Xu and Yang (2012) that additionally corrects for variability bias. We choose here to correct only the mean bias, specifically to reduce the strong wind shear. Although the variance in CCSM′ is generally slightly smaller than NNRP′ over the base period, the regional model driven by CCSM R recovers this difference in variance (not shown). Equation 3 is applied to variables needed to generate the lateral boundary conditions for NRCM; zonal and meridional wind, geopotential height, temperature, relative humidity, mean sea level pressure, and lower boundary condition of sea surface temperature. When driven by revised boundary conditions, the NRCM develops substantially improved wind shear over the tropical Atlantic (Fig. 3).

There may be sensitivity of the revised climate to the choice of the base period arising from non-stationarity of the bias, but fortunately this is not the case here. Sensitivity studies using different base periods result in nearly identical bias corrections over the entire simulation period. This increases confidence that the bias will not change substantially in the future, though the validity of this assumption needs further consideration. Further justification for this approach is provided by Mote et al. (2011), who showed that for assessing changes between a current and future period, a biased model for current climate is as valid as an unbiased model. On the other hand, Dosio and Paruolo (2011) showed the assumption of constant bias in time may only be appropriate for ensemble mean quantities rather than for individual model projections.

2.3 Future changes in tropical cyclone frequency

NRCM simulated tropical cyclones are identified and tracked automatically following Suzuki-Parker (2012) and described briefly in the SOM section S2. On the 36 km domain the NRCM driven by revised CCSM data produces an average of 7.6 North Atlantic tropical cyclones annually for base climate. As noted earlier the model base climate period 1995–2005 is nominal. This introduces ambiguity as to the selection of observational period used for model evaluation. The selection of 1975–1994 as base period for the bias correction leads to this being one logical choice. In this case the comparative annual number of observed North Atlantic named tropical cyclones (including tropical storms and hurricane categories 1–5) is 8.9 (using IBTrACS data, Knapp et al. 2010). Alternative choices are the average over, say, the recent 50 years (1958–2007) of 10.4, or the current (1995–2005) high period of 14.3 cyclones per year. Regardless of the choice of observational period for comparison, NRCM appears to underestimate the actual frequency. This frequency underestimate could easily be corrected by tuning the detection criteria but the detection parameters are deliberately frozen at fixed values. Of importance to this study, Suzuki-Parker (2012) showed that the future changes in cyclone frequency are not impacted by the choice of detection values.

The NRCM predicts an increase in North Atlantic tropical cyclone frequency, with annual numbers on the 36 km domain increasing from 7.6 in the base climate to 8.5 in 2020–2030 and 10.4 in 2045–2055. This is a statistically significant increase (using the Wilcoxon-Mann–Whitney rank-sum test, Wilks (2006)) at the 90 % level and represents a 37 % increase in North Atlantic tropical cyclone frequency over the next 50 years. Annual numbers of cyclones on the 12 km domain, using the same detection criteria as for 36 km, are higher at 17.3, 17.5 and 19.9 for the three time slices, but show a similar increase (though not significant) of 15 % over the next 50 years. These results are different from those of other studies, which have tended towards predicting small changes and if anything a decrease in overall Atlantic tropical cyclone frequency over coming decades (e.g. Knutson et al. 2010; Bengtsson et al. 2007), and serves to highlight the large uncertainty in determining changes in high-impact events on regional scales.

The modeled spatial distribution of North Atlantic tropical cyclone track and genesis densities on both the 36- and 12 km domains for base climate are close to the observed long-term climatology (Fig. 4). The future climate prediction exhibits a consistent southeastward shift in track density, from a maximum in the mid-Atlantic in base climate, to a low-latitude maximum in 2045–2055 (Fig. 4). This prediction is a continuation of recent trends (Holland and Webster 2007) and agrees with Wu et al. (2010) who found that a relative increase in eastern Atlantic SSTs leads to changes in atmospheric circulation and near-equatorial tropical cyclone activity.

Fig. 4
figure 4

Tropical cyclone track density (color shading) and genesis density (contours) normalized by the maximum value for (top) IBTrACS data 1975–1995, (left column) NRCM 36 km domain, (right column) NRCM 12 km domain, (top row) base climate, (middle row) 2020–2030, and, (bottom row) 2045–2055. Density is defined as the number of cyclone tracks or genesis points within a 5 degree radius of a point per year. Non-normalized track and genesis densities are provided in the SOM section S4

2.4 Future changes in tropical cyclone intensity

Modeled tropical cyclone intensity distributions show the most intense cyclones are weaker than observed (Fig. 5). This is a common problem due to the inability of the 36- and 12 km grids to resolve the inner core dynamics of a tropical cyclone (e.g. Knutson et al. 2008; Davis et al. 2010; Gentry and Lackmann 2010). Note also the related tendency to over-simulate moderate intensity systems, which to some extent is due to those storms that would have been more intense being held back in this region. Nevertheless, storms on the 36 km and 12 km domains experience modest future increases in mean wind speed of approximately 0.64 ms−1 and 0.90 ms−1 respectively. This increase is less than observational errors in the historical record; meaning current observation systems could not detect this change in the mean intensity.

Fig. 5
figure 5

Frequency distributions of 6-hourly tropical cyclone maximum wind speed (ms−1) at 10 m above the surface as simulated by the NRCM (top) 36 km domain and (bottom) 12 km domain for base climate (black line), the period 2020–2030 (dark gray), and the period 2045–2055 (light gray), and observations (IBTrACS data) for the period 1995–2005 (black dashed line). Adapted from Holland et al (2010). Copyright 2010 OTC. Reproduced with permission of the copyright owner

A more marked increase, however, is seen in the number and intensity of the most intense hurricanes that can be resolved by the model (Fig. 5). This is in agreement with other dynamical modeling and theoretical studies (Knutson et al. 2010) and for recent changes to current climate (Holland and Bruyère 2013). These intensity increases may be due to future large-scale environment changes (indeed, Suzuki-Parker (2012) showed favorable future increases in SST, relative SST (Vecchi and Soden (2007) and decreasing shear in the driving CCSM data) but may also be due to the future southeastward shift of track density (Fig. 4) associated with potential changes in the proportion of tropical cyclones developing from easterly waves and cyclone track lengths.

3 Statistical assessments

Confidence in the variability and trends of high-impact events obtained through dynamical downscaling is limited in part by the relatively short period and small number of events that can be simulated. Fortunately, sample size can be increased substantially through statistical approaches and can aid uncertainty assessments. Such statistical assessments provide a valuable adjunct to the dynamical simulations and by way of illustration two approaches are discussed here.

3.1 Empirical assessments of tropical cyclone frequency

A popular statistical downscaling approach for tropical cyclone frequency is the use of empirical relationships with large-scale data taken from reanalysis or global model data (e.g. Gray 1968, 1984; Emanuel and Nolan 2004). These approaches were typically designed to capture the hemispheric seasonal cycle and typically do not do well on regional and interannual scales. For example, Menkes et al. (2011) applied four genesis potential indices to different reanalysis datasets and found poor reproduction of interannual amplitude and phase variability on regional scales. However, recent work has begun to show skill on regional, seasonal scales.

The Emanuel and Nolan (2004) index combines low-level vorticity, mid-level relative humidity, potential intensity (a measure of the vertical instability of the atmosphere, Emanuel 2000), and vertical wind shear. Bruyère et al. (2012) showed that for interannual variability and longer-term changes, the relative humidity and vorticity terms contribute nothing to the skill, though this could be due to the specific formulation of the index rather than having a physical interpretation (a general limitation highlighted by Menkes et al. 2011). Bruyère et al. (2012) also found care needs to be taken when selecting an index averaging area: for the North Atlantic, a basin-wide average was not optimal in explaining total basin cyclone frequency, whereas an average taken over the eastern tropical Atlantic (5–20oN, 60–15oW), was able to explain 72 % of the annual variance of total basin cyclone frequency.

Suzuki-Parker (2012) applied the Bruyère et al. (2012) index to the same revised global data used to drive the NRCM simulations presented in Section 2 and produced results in agreement with the NRCM dynamical results of an increase in tropical cyclone frequency of between 1 and 3 storms by the mid 21st century.

3.2 Extreme value assessment of tropical cyclone intensity

One complimentary approach to assess changes to the most intense cyclones from the truncated regional climate simulations is to utilize extreme value statistics (Coles 2001). The NRCM-generated cyclone intensity distribution is truncated at maximum winds of around 43 ms−1 for tropical cyclones on the 36 km grid (Fig. 5). However, as stated earlier the NRCM predicts a shift of the truncated distribution towards more intense storms over the next 50 years (Fig. 5). Applications of the Weibull distribution (Weibull 1939) have been experimented with to assess associated changes in the intense cyclones. The Weibull is a stretched exponential distribution and is used here because of its history of application to modeling weather and climate extremes (e.g. Katz and Brown 1992; Mearns et al. 1984) and because it is bounded below at zero.

The Weibull Cumulative Distribution Function (CDF) and Probability Density Function (PDF) are:

$$ CDF:\kern0.5em F(x)=1-{e}^{-{\left(\frac{x}{\alpha}\right)}^{\ss }},\kern0.5em 0\le x\le \infty $$
(4)
$$ PDF:\kern0.5em f(x)=\frac{\partial F(x)}{\partial x}=\frac{\ss }{\alpha }{\left(\frac{x}{\alpha}\right)}^{\ss -1}\ {e}^{-{\left(\frac{x}{\alpha}\right)}^{\ss }},\kern0.5em 0<x<\infty $$
(5)

The scale parameter α and the shape parameter β both lie in the range (0,∞): β = 1 corresponds to the exponential distribution, β = 2 the Rayleigh distribution, and β = 3.5 an approximation of the normal distribution. The mean and the variance of the Weibull distribution are given by:

$$ Mean:\kern0.5em \mu =\alpha \varGamma \left(1+\frac{1}{\ss}\right) $$
(6)
$$ Variance:\kern0.5em {\sigma}^2={a}^2\varGamma \left(1+\frac{2}{\ss}\right)-{\mu}^2, $$
(7)

where Γ is the Gamma function. Thus, a first-order approximation to future changes in tropical cyclone extremes can be obtained by first applying the Weibull to current tropical cyclones for the period 1995–2008 (where x in Eqs. 4 and 5 corresponds to cyclone lifetime maximum 10 m wind speed, using IBTrACS, Knapp et al. 2010) to obtain current scale parameter α =37 and shape parameter β =1.8 using the method of moments. Model-resolved future changes in the mean and standard deviation of tropical cyclone intensity are then used to estimate changes to these current scale and shape parameters and provide an assessment of future full intensity distributions. Finally, the changes in probability of various tropical cyclone intensities can be calculated from the Weibull exceedence probability:

$$ P\left(x>c\right)=1-F(c)={e}^{-{\left(\frac{c}{\alpha}\right)}^{\ss }}, $$
(8)

where c is the lower limit of the intensity range of interest (e.g. 69 ms−1 for Cat 5 hurricanes).

The small 36 km NRCM-predicted changes in the resolved distribution result in a much greater change in the most intense cyclones (Fig. 6). Category 5 hurricanes are predicted to increase by 60 % from a base climate period of 1980–1994 and by 30 % from a base climate period of 1995–2008. In this application we make the assumptions that the modeled changes are indicative of the changes that would have been simulated in the full distribution, were it resolved; and there is no process that will cause a change in the unresolved tail of the distribution without any signal in the resolved component.

Fig. 6
figure 6

Future changes to all hurricanes (green), category 3–5 (yellow), category 4–5 (red) and category 5 (blue), using a base climate of: 1980–1994 (left) and 1995–2008 (right)

Another approach is to further dynamically downscale the NRCM to a resolution capable of resolving the full intensity distribution. Bender et al. (2010) took this approach using the Geophysical Fluid Dynamics Laboratory operational hurricane model to downscale to a resolution that captured the full intensity distribution and showed similar future changes to the most intense storms as reported here. This suggests that the statistical correction of the intensity distribution is a promising line of research that warrants further exploration, particularly with regard to understanding how the model intensity distribution relates to the observed intensity distribution.

4 Concluding discussion

This discussion of dynamical and statistical approaches to assessing variability and change in high-impact weather events is intended to contribute to community discussion and collaboration on best practices. Key findings are:

  1. 1)

    Importance of domain size, location and resolution: These aspects of model setup need careful consideration in order to capture accurate regional climate and high-impact weather. Experience with the NRCM using large domains at different resolutions indicates that: higher resolution than is required to resolve the key characteristics of the high-impact events may produce little added value and may not be an optimal use of resources; and, to capture accurate regional climate and regional climate change the regional domain may need to include those forcings and circulations that directly affect the regional climate over the area of interest.

  2. 2)

    Treating bias: Regional climate simulations can be severely affected by biases in the driving global climate model, even when large domains are employed. A successful technique has been shown to remove such bias in a manner that retains the day-to-day weather, climate variability and change components. As climate models improve, there remains a need to treat regional biases (e.g. Muñoz et al. 2012). Further work is required to fully understand the implications of such techniques, to assess and remove changes in bias with time, and to develop new approaches to bias removal for emerging modeling tools such as regional atmosphere-ocean coupled models and global variable resolution meshes.

  3. 3)

    Incorporating statistics and assessing uncertainty: Confidence in high-impact weather assessments obtained through dynamical downscaling is limited by both truncation of the observed distribution and the small number of events that can be simulated. Ensemble modeling is an obvious approach to increase sample size and explore statistical uncertainty. Indeed, Strachan et al (2013) show spread in annual TC counts in a small initial condition ensemble global atmospheric model, but this approach is computationally impractical for high resolution, large domains. The examples of empirical and extreme value approaches discussed here serve to demonstrate the high potential of combining dynamical modeling with statistical approaches in assessing high-impact weather. Their value lies in both assessments of events beyond the capacity of current dynamical models and the effective increase in sample size by the ability to use low-resolution ensembles. These lead to improved assessments of statistical uncertainty in the climatology of high-impact weather and future change.

Meeting the societal demand for assessments of variability and change in high-impact weather events with regional clarity is a challenge that extends well beyond simply improving climate simulations, and once the regional climate prediction has been made and the uncertainty quantified, there remains a huge gap regarding how this affects society. The past two decades of regional climate research outputs have largely been exploratory in nature and not directly aligned to the requirements of the end user. Exceptions include ensemble regional downscaling programs and bias corrected and statistically downscaled CMIP3 archives. Recent work by Towler et al (2012) went a step further by incorporating NRCM-generated climate change data into a risk-based approach to assess ecological impacts and inform conservation efforts. The field is wide open to develop the concepts discussed here further and to more fully integrate with societal impact assessments.