North American extreme temperature events and related large scale meteorological patterns: a review of statistical methods, dynamics, modeling, and trends
- First Online:
- Cite this article as:
- Grotjahn, R., Black, R., Leung, R. et al. Clim Dyn (2016) 46: 1151. doi:10.1007/s00382-015-2638-6
The objective of this paper is to review statistical methods, dynamics, modeling efforts, and trends related to temperature extremes, with a focus upon extreme events of short duration that affect parts of North America. These events are associated with large scale meteorological patterns (LSMPs). The statistics, dynamics, and modeling sections of this paper are written to be autonomous and so can be read separately. Methods to define extreme events statistics and to identify and connect LSMPs to extreme temperature events are presented. Recent advances in statistical techniques connect LSMPs to extreme temperatures through appropriately defined covariates that supplement more straightforward analyses. Various LSMPs, ranging from synoptic to planetary scale structures, are associated with extreme temperature events. Current knowledge about the synoptics and the dynamical mechanisms leading to the associated LSMPs is incomplete. Systematic studies of: the physics of LSMP life cycles, comprehensive model assessment of LSMP-extreme temperature event linkages, and LSMP properties are needed. Generally, climate models capture observed properties of heat waves and cold air outbreaks with some fidelity. However they overestimate warm wave frequency and underestimate cold air outbreak frequency, and underestimate the collective influence of low-frequency modes on temperature extremes. Modeling studies have identified the impact of large-scale circulation anomalies and land–atmosphere interactions on changes in extreme temperatures. However, few studies have examined changes in LSMPs to more specifically understand the role of LSMPs on past and future extreme temperature changes. Even though LSMPs are resolvable by global and regional climate models, they are not necessarily well simulated. The paper concludes with unresolved issues and research questions.
KeywordsLarge scale meteorological patterns for temperature extremes Heat waves Hot spells Cold air outbreaks Cold spells Statistics of temperature extremes Dynamics of heat waves Dynamics of cold air outbreaks Dynamical modeling of temperature extremes Statistical modeling of extremes Trends in temperature extremes
1 Introduction to temperature extremes
Temperature extremes have large societal and economic consequences. While many heat waves are short-lived, longer events can have a large economic cost. Cold air outbreaks (CAOs) tend to be short-lived but carry large economic losses. Timing of the CAOs can be more important than the minimum temperatures of the freeze; during 4–10 April 2007 low temperatures across the South caused $2B in agricultural losses since many crops were in bloom or had frost sensitive buds or nascent fruit (Gu et al. 2008). The event also exemplifies how monthly means can be misleading: April 2007 average temperatures were near normal. In short, both hot spells (HSs) and CAOs have great societal importance and they are short-term events that do not necessarily appear in monthly mean data.
This report focuses on short-term (5-day or less) extreme temperature events occurring in some part of North America. Temperature extremes considered in this paper include both short-term hottest days (warm season) and CAOs (winter and spring) as these have the greatest impacts. Such events, both observed and simulated, have received considerable attention (including research papers, e.g. Meehl and Tebaldi 2004; and active websites: http://www.esrl.noaa.gov/psd/ipcc/extremes/, http://www.ncdc.noaa.gov/extremes/cei/, and http://gmao.gsfc.nasa.gov/research/subseasonal/atlas/Extremes.html). However, the emphasis here is on the less well understood context for the extreme events. Our primary context is the large-scale meteorological patterns (LSMPs) that accompany these extreme events.
Temperature extreme events are usually linked to large displacements of air masses that create a large amplitude wave pattern (here called an LSMP). LSMPs have a spatial scale bigger than mesoscale systems but smaller than the near-global scale of some modes of climate variability. The LSMP often has some portion that is superficially similar to a blocking ridge, so a blocking index can be an LSMP indicator (Sillmann et al. 2011). Extreme events have also been linked to circulation indices like the North Atlantic Oscillation (NAO) (Downton and Miller 1993; Cellitti et al. 2006; Brown et al. 2008; Kenyon and Hegerl 2008; Guirguis et al. 2011) the Madden–Julian Oscillation (MJO) (Moon et al. 2011) and El Niño/Southern Oscillation (ENSO) (Downton and Miller 1993; Higgins et al. 2002; Carrera et al. 2004; Meehl et al. 2007; Goto-Maeda et al. 2008; Kenyon and Hegerl 2008; Alexander et al. 2009; Lim and Schubert 2011). However, LSMPs are likely distinct from climate modes for several reasons. First, named climate modes such as the NAO are common modes of variability, whereas the LSMP is presumably as rare as the associated extreme event. Second, climate modes occur on a longer time scale than the LSMPs for the short-term events focused upon in this article. However, it is possible that an extreme event might occur when a climate mode has transient extreme magnitude or is amplified in association with another low frequency phenomenon. Third, tested LSMP patterns are not that similar to climate modes. In correlating eight NOAA teleconnection patterns (http://www.cpc.ncep.noaa.gov/data/teledoc/telecontents.shtml) and California LSMPs and in assessing the PNA contribution to last winter’s extreme cold in eastern North America (neither shown here), we do not find notable contribution from such modes. Several studies identified the LSMPs associated with specific extreme hot events (Grotjahn and Faure 2008; Loikith and Broccoli 2012) and CAO events (Konrad 1996; Carrera et al. 2004; Grotjahn and Faure 2008; Loikith and Broccoli 2012). Parts of these LSMPs tend to be uniquely associated with the corresponding extreme weather and those parts have some predictability (Grotjahn 2011).
The LSMPs for extreme events are not fully understood for different parts of North America. Also, local processes: topography, soil moisture, etc. play key roles but there is a knowledge gap in how well climate models simulate the LSMPs as well as these local processes and how the local and global modes interact with LSMPs. Bridging these knowledge gaps will reduce the uncertainty of future projections and drive model improvements.
Now is an opportune time to summarize critical issues and key gaps in understanding temperature extremes variability and trends because: (1) it is not known if current climate models used for future projections are producing extremes via the correct dynamical mechanisms, which directly impacts confidence in projections, and (2) knowledge of the LSMPs can improve downscaling (statistical or dynamical) by focusing attention on large scale patterns that are fundamental to the occurrence of the extreme event. Conversely, global models that do not reproduce the magnitude or duration of extreme temperature events accurately may still capture the correct LSMPs and facilitate downscaling. Finally, there is now sufficient preliminary work and growing interest to make a summary valuable.
From the LSMP context, the objectives of this review are surveys of: relevant statistical tools for extreme value analysis (Sect. 2), synoptic and dynamical interactions between LSMPs and other scales from local to global (Sect. 3), model simulation issues (Sect. 4), trends in these temperature extremes (Sect. 5), and various open questions (summary section). Different readers may be interested in different surveys. Accordingly, the sections are autonomous allowing a reader to skip to a particular section(s) of interest.
2 Extreme statistics and associated large scale meteorological patterns (LSMPs)
2.1 Definitions of extreme events
Some temperature related ETCCDI indices (Sillmann et al. 2013a)
ETCCDI index name
The percentage of days when the high temperature is greater than 90 % of those in reference period
The percentage of days when the high temperature is less than 10 % of those in reference period
The percentage of days when the low temperature is greater than 90 % of those in reference period
The percentage of days when the low temperature is less than 10 % of those in reference period
Monthly or seasonal maximum of daily maximum temperature
Monthly or seasonal minimum of daily maximum temperature
Monthly or seasonal maximum of daily minimum temperature
Monthly or seasonal minimum of daily minimum temperature
Heat Wave Duration Index
Length of a heat wave
Cold Wave Duration Index
Length of a cold spell
Days below freezing
2.2 Application of extreme value statistical techniques
The observed changes in Fig. 1 are calculated using a time-dependent point process approach to fit “peaks over threshold” statistical models (Coles 2001). In this case, the extension of stationary extreme value methods to a time-dependent formalism used time as a “covariate” quantity to the ETCCDI indices (Kharin et al. 2013). A principal advantage of a fully time dependent formalism over a quasi-stationary approximation (Wehner 2004) is that the amount of data used to calculate extreme value parameters is substantially increased, resulting in higher quality fitted distributions and hence more accurate estimates of long period return values. Calculations involving climate model output gain additional statistical accuracy by using multiple realizations from ensembles of simulations, provided they are independent and identically distributed.
The “block maxima” and “peaks over threshold” methods to fit the tails of the distributions of random variables are asymptotic formalisms (Coles 2001). In this terminology, “block maxima” refers to use of only the maximum value during each “block” of time, usually a single season or year. The resulting generalized extreme value (GEV) or Poisson and generalized Pareto distributions (GPDs) are both three parameter functions and can be transformed between each other. Hence provided that the data used to fit a distribution are in the “asymptotic regime”, i.e. far out in the tail of the distribution, the two methods are equivalent. Uncertainty in the estimate of long period temperature return values resulting from limited sample size can be appreciable and may be as large as that from unforced internal variability (Wehner 2010). However, variations in these estimates from multi-model datasets such as CMIP3/5 are generally significantly larger.
Confidence in the estimates of the statistical properties of the tail of the parent distribution of a random variable can be ascertained by exploring the sensitivity to the sample size used to fit the extreme value distribution. For block maxima methods, the length of the block is a season (or effectively so for temperature if the block length is a year). Lengthening the block makes the sample size smaller; shortening it makes the sample size larger since only one extreme value is drawn from each block. Such sensitivities are somewhat more straightforward to explore with peaks over threshold (POT) methods. Typical thresholds may be chosen between 80 and 99 % depending on the size of the parent distribution that the extreme values are drawn from. However, standard POT methods may not discriminate between extreme values that occur at successive dates when the individual extreme values may not be truly independent. In these cases, declustering techniques (Coles 2001) are applied to avoid biased (low) estimates of the uncertainty. The trends in extreme temperature shown in Fig. 1 are calculated using such a declustering technique and a POT formalism with time as a linear covariate. An alternative approach retains possibly dependent consecutive extremes, adjusting the estimates of uncertainty through either resampling or more advanced techniques for quantifying extremal dependence (e.g., Fawcett and Walshaw 2012). Furthermore, LSMPs responsible for extreme events can be formed using only the dates of the onset of the event (e.g. Grotjahn and Faure 2008; Bumbaco et al. 2013) reducing the risk of autocorrelated extremes. In the cited studies, a 5 day gap was typically required between events. Low frequency factors (or climate modes), such as ENSO, are best treated using the previously mentioned covariate techniques.
Some of the advantages and challenges of applying statistical methods based on extreme value theory to analyze non-stationary climate extremes have been pointed out previously (e.g., Katz 2010), but are still not necessarily well appreciated by the climate science community. Conventional approaches tend to be either: (1) less informative (e.g., analyzing only the frequency of exceeding a high threshold, not the excess over the threshold and not measuring the intensity of the event); or (2) less realistic (e.g., based on assumed distributions such as the normal that may fit the overall data well, but not necessarily the tails). When relating extremes to LSMPs, standard regression approaches would not quantify the uncertainty in the relationship as realistically as using extremal distributions with covariates. Challenges in extreme value methods include specifying the dependence on LSMPs of the parameters of the extremal distributions in a manner consistent with our dynamical understanding. Moreover, heat waves and CAOs are relatively complex forms of extreme events, some of whose characteristics can be challenging to incorporate into the framework of extreme value statistics (Furrer et al. 2010). Finally, another advantage of the POT approach over the block maxima approach is being able to incorporate daily indices of LSMPs as covariates (not just monthly or seasonally aggregated indices).
2.3 Identification of LSMPs related to extreme temperatures
A summary of various methodologies used to identify large scale meteorological patterns (LSMPs)
Method (example reference)
Composites (e.g. Grotjahn and Faure 2008)
Average together an ‘ensemble’ of maps on pre-identified dates As with some other procedures described here, it is important to assess how independent the dates are, e.g. consulting the autocorrelation
Ensemble average provides physical insights. Can be used to examine patterns leading up to and subsequent to the event onset by shifting dates used. Non-parametric in making no assumption about the pattern. The dates can be chosen to be independent
Must identify the event dates beforehand. Only finds one average unless individual members of ensemble examined and grouped when choosing which dates are in which average
Bootstrap resampling can compare ensemble average to many random ensemble averages (e.g. Grotjahn and Faure 2008). Need larger sample size or another method to assess consistency (e.g. sign counts: Grotjahn 2011)
Regression (e.g. Lau and Nath 2012)
Fit a regression line (polynomial in the predictand) to a time series of a predictor at each grid point
Pattern provides physical insights
Parametric in assuming a specific regression line (polynomial). Can be used to examine patterns leading up to dates chosen event if event lasts for one time sample
Fit of regression line (polynomial) to extreme values may be notably altered by the choice of the polynomial assumed. Only finds one pattern. Does not incorporate time threshold criteria (e.g. event must last >2 days) Treats all dates as independent, which they might not be. Does not separate onset from other days during an event (e.g. mixing onset with dates during the event). However, low pass filtering sometimes used to aggregate the mixture of onset and during-event dates
Significance can be estimated by rejecting a null hypothesis about the coefficients. based on where the regression line values are higher (or lower) than a specified threshold of the predictand values
EOFs or PCs (Wu et al. 2012)
Calculate the eigenvectors of a space-weighted covariance (between values at different grid points) matrix. If only inputting extreme dates, then it is important to assess how independent the dates are, e.g. consulting the autocorrelation
Finds multiple patterns, each of which is orthogonal (not a subset) of another pattern. Can identify fraction of variability that is due to that EOF/PC. Most often used with filtered data to find leading low frequency structures
Not suitable when applied to all data as the leading EOFs will be common patterns not necessarily relevant to rare events. EOFs more useful when calculated only from data on event dates identified beforehand. Eigenvectors may not be sufficiently distinct. The patterns found may depend on domain chosen, though EOF ‘rotation’ may help. Each EOF explains only a fraction of the variance and so no single EOF might be an LSMP thereby limiting physical insight
No inherent significance test since the EOFs (or PCs) may not represent an actual weather pattern. The amount of variance associated with a particular EOF/PC is often used to indicate the importance of that EOF/PC
SOMs (e.g. Hewitson and Crane 2002)
Obtain the distribution of patterns covered by a set of input maps, using neural-net training
Resultant maps span the pattern space of input fields, represent nodes of a continuous pattern distribution, and can be related to a transformation of rotated EOFs back to physical space
Results can be sensitive to the domain used for pattern analysis. Care needed to ensure appropriate resolution of LSMP space. Low resolution can fail to resolve pattern features important for extreme events/Overly fine resolution undermines significance testing
Bootstrapping methods can compare frequency distributions of extreme-event days in LSMP space with random sampling of all days to indicate significance of event clustering in subregions of the full pattern space. (e.g. Cassano et al. 2007)
Clustering analysis (e.g. Stephanon et al. 2012)
Assigning events to K clusters so that it minimizes total sum of the generalized Euclidian distance between patterns in a cluster (Spath 1985; Seber 2008). If only inputting extreme dates, then it is important to assess how independent the dates are, e.g. consulting the autocorrelation
Objectively classify events and find relating LSMPs that provide physical insight
No assumptions about the patterns such as orthogonality and symmetry
The number K of LSMP clusters can be somewhat arbitrary since K is pre-specified, but the ‘dissimilarity index’ can estimate the separation between clusters for different K and help optimize the value of K. Larger separation between clusters is preferred and generally declines as K increases. Optimal K can be unclear for some situations.
Significance of classification stability can be obtained from Monte Carlo test by rejecting the null hypothesis that the verification period temperature extreme cannot be classified in the clusters obtained from remaining periods. Lee and Grotjahn (2015) use bootstrap method (random ensembles same number of members as the cluster) to identify what parts of each cluster average are notable for the extreme event
Machine learning (Salakhutdinov and Hinton 2012)
Train a multi-layer neural network on a dataset. Training is done one layer at a time, with spatio-temporal patches at the bottom layer, and class labels being assigned at the top layer
Deep Belief Networks have proved to be powerful in capturing a range of patterns. It is likely that they will extract invariant patterns as well as anomalies
These techniques have produced state-of-the-art results in computer vision and speech recognition tasks. They have not yet been applied to climate datasets
Has not been conducted in the context of climate/LSMP applications. Classification performance of method has been conducted extensively for images by testing on held-out data
Sampling of Various Criteria Used in Heat Wave and Hottest Day Definitions
A period of at least 48 h during which neither the overnight low nor the daytime heat index Hi falls below the NWS heat stress thresholds (80 and 105 °F). At stations where more than 1 % of both the high and low Hi observations exceed these thresholds, the 1 % values are used as the heat wave thresholds
Hajat et al. (2002)
3-day moving average temperatures >the 99th percentile of the whole record of temperature
Meehl and Tebaldi (2004)
The longest period of consecutive days satisfying the following 3 conditions
1. Daily maximum temperature >T1 for at least 3 days
2. Average daily maximum temperature >T1 for entire period
3. Daily maximum temperature >T2 for every day of entire period,
T1 (threshold 1) = 97.5th percentile of distribution of maximum temperatures in the observations and in simulated present day climate
T2 = 81st percentile
Maximum T exceeding the 90th quantile of summer temperature (30 °C) at a station (Basil, Switzerland)
Lipton et al. (2005)
Daily maximum high temperature remains 2 standard deviations above normal for at least 2 consecutive days
Gosling et al. (2007)
For 3 or more days the maximum T must be ≥95th percentile of the maximum T in the summer climatology
Grotjahn and Faure (2008)
At least 3 consecutive days during which the daily maximum temperatures are above 100 °F (38 °C), and with at least one above 105 °F (40.5 °C)
Two combinations of criteria, I and II, were tested
I. Must satisfy both conditions (where anomaly is relative to long term daily mean):
1. At least 3 consecutive days with daily anomaly maximum temperature ≥10 °C
2. At least 1 day must have maximum temperature anomaly ≥15 °C
II. Must satisfy the 2 conditions above plus this additional condition:
3. The average maximum temperature for the event ≥100 F (38 °C)
Gershunov et al. (2009)
Individual stations exceeding the 99th percentile for 1, 2, or 3 dates in a row are aggregated, with the highest aggregation of values over the region including all of California and Nevada determining a ranking for an event. Daytime maximum and nighttime (highest) minimum treated separately
Daily maximum temperature must exceed the 90th percentile for at least 3 consecutive days, where the percentile is based daily values from the 3-month summer season. Also tested, same temperature criterion over 5 consecutive days
Daily maximum temperature anomaly (relative to long term daily mean) normalized by daily long term mean standard deviation at all three CV stations (KRBL, KFAT, KBFL) must all exceed 1.6 Note: this defines hottest days, not heat waves
Bumbaco et al. (2013)
Daily maximum temperature anomalies for stations in a region are averaged together. Heat wave when regional average daily anomaly exceeds 99th percentile for 3 or more consecutive days
Compositing has some disadvantages that can be addressed. The identification of the extreme event dates must be done separately and before the compositing. The composite produces one target ensemble average (for a specific field and level) for each set of target dates. If more than one LSMP can produce the extreme event, then that must be identified either with one or more additional criteria when choosing dates or identified by examining (perhaps qualitatively) the maps for each individual member of the target ensemble. A procedure like adding the number of positive and subtracting the number of negative anomaly values at a grid point in the target ensemble members (called the ‘sign count’ in Grotjahn 2011) can assist with identifying multiple patterns. (If the sign count equals the number of ensemble members, then all members have the same sign of the anomaly field at that location.) Lee and Grotjahn (2015) apply a cluster analysis to distinctly different parts of LSMPs prior to California heat waves and identify two ways the onset LSMPs form.
Regression estimates one quantity (the predictand) using a function of one or more other quantities (the predictors). The method is often parametric in assuming a specific function relates a predictor to the predictand (but nonparametric methods exist, too). An example predictor might be daily minimum surface temperature and the predictand might be 700 hPa level meridional wind. At each grid point the value of the predictand can be estimated using a polynomial function of the predictor, where the coefficients of that polynomial are calculated to minimize a squared difference between actual predictand values and polynomial values at that grid point. In general, the coefficients differ from grid point to grid point. To find the LSMP in this example of extreme cold events, the polynomial can be used to construct the predictand at each grid point using a predictor value such as two standard deviations below normal.
The pattern obtained by regression can provide physical insights by directly linking the patterns in the predictand to extremes in the predictor. For example, lower tropospheric (700 hPa) winds prior to a California CAO flow from northern Alaska and northern western Canada to reach California without crossing over the Pacific Ocean. Regression can be used to examine patterns leading up to (or after) event onset dates by offsetting in time the predictor values from the predictand values when calculating the regression coefficients. The LSMP is again the resultant predictand when the predictor is at some specified value (e.g. predictor equals two standard deviations below normal). Significance can be estimated by rejecting a null hypothesis (e.g. that the regression coefficient is zero at the 1 % level using a student’s t test).
One disadvantage of regression is that the assumption of a specific polynomial to represent how the predictand varies relative to the predictor. The fit of the regression line (polynomial) to extreme values may be notably altered by the order of the polynomial assumed. Regression, like composites, only finds one pattern. Regression does not incorporate event duration criteria (such as: the event must last at least 3 days). Regression treats all dates as independent, which they are not likely to be; however this can be somewhat mitigated by sub-sampling the data (e.g. only use every fifth day). Subsampling might be combined with low pass filtering to aggregate the mixture of onset and during-event dates. Another disadvantage is that a portion of the pattern may be highly significant but only have small correlation to the predictor.
2.3.3 Empirical orthogonal functions
Empirical orthogonal functions (EOFs) or principal components (PCs) can be used to identify LSMPs for extreme events. EOFs are the eigenfunctions of a matrix formed from the covariance between grid points on maps. EOFs from all maps in a time record will be ordered based on the eigenfunctions responsible for the largest amount of variance between time samples. Such eigenfunctions are the most common modes of variability and so not likely to be LSMPs of extreme events that are rare (except as mentioned in the introduction). However, EOFs can be formed only from maps selected in reference to an extreme event, such as maps only on the target dates of events onset. (EOFs of common low frequency modes may influence short-term extreme events focused on in this paper, as discussed in Sect. 3.1.) Weighting can be used for a variable grid spacing (such as occurs when using equal intervals of longitude across a range of latitudes).
An advantage of EOFs/PCs is the method finds multiple patterns, each of which is orthogonal (not a subset) of another pattern. The method can calculate the fraction of variability that is due to each EOF/PC. This approach is most often used with filtered data to find low frequency structures. This method is suitable for finding patterns leading up to and after the event by shifting the dates chosen by the event criteria (and only using those dates). In a study of California Central Valley hot spells, Grotjahn (2011) found the leading EOF based on dates satisfying the criteria in Table 3 to be very similar to the corresponding ensemble mean composite.
A disadvantage is that the patterns found may depend on the domain chosen, though EOF ‘rotation’ may help. Also, each EOF may explain only a small fraction of the variance and no single EOF might be an LSMP thereby limiting physical insight. Different leading EOFs/PCs might have structures influenced by the required orthogonality and possibly not a pattern that occurs during an event. While the amount of variance associated with a given EOF is used to indicate the importance of the EOF, there is no inherent statistical significance test. Hence it can be unclear what portions of the EOF are significantly associated with the extreme event and what parts are not (and happen to reflect limited variation in the finite sample).
2.3.4 Clustering analysis
Clustering analysis is terminology indicating a widely used partitioning procedure that identifies separate groups of objects having common structural elements. Clustering analysis has been used to classify distinct sets of LSMPs associated with extreme events (Park et al. 2011; Stephanon et al. 2012). When we have 100 historical hot spell events over a given region, not all extreme events may have the same LSMP on or prior to their onset. Some events may have a wave-like height field, others may have a dipole pattern, and still others may have a third pattern. Although detailed grouping procedures vary for every clustering technique, the basic concept is to minimize the overall distance between patterns among events in resultant groups. For example, the k-means clustering technique applies an iterative algorithm in which events are moved from one group to another until there is no additional improvement in minimizing the squared Euclidean point-to-centroid distance in a group (Spath 1985; Seber 2008), where each centroid is the mean of the patterns in its cluster.
Output of clustering analysis is just the average field of events in each cluster, similar to the output from composite analysis. Unlike composite analysis in which members of clusters are pre-identified, the essential point of clustering analysis is to objectively classify events based on spatial pattern similarity. By applying cluster analysis to group similar onset patterns, one can isolate distinct dynamical origins of different extreme temperature events. Another advantage is that resultant clusters are based on physical maps without assumptions of orthogonality and symmetry such as in the mode separation by EOF/REOF. The robustness of a hot spell classification can be tested by a Monte Carlo test as follows. First one calculates a stability score that is the ratio of the number of verification period heat waves that are correctly attributed over the total number heat waves in the cluster. Second one estimates the probability density function (PDeF) of the null hypothesis that cluster assignment is purely random. Significance of each cluster can be estimated by rejecting a null hypothesis (e.g. stability score is located within highest 99 % of PDeF).
A disadvantage of clustering analysis is that one pre-specifies the number of clusters (e.g. k in k-means clustering). Determining the number k is subjective if one does not have sufficient prior knowledge of related physical patterns. There are several statistics to check the optimal number of k such as ‘distance of dissimilarity’ (Stephanon et al. 2012). Another disadvantage is the ambiguity of cluster assignment for certain events. Another disadvantage is the patterns of such events cannot be assigned clearly to one group over another group. Part of a pattern may resemble Cluster #1, while another part may be more similar to Cluster #2. To avoid ascribing some marginal events to specific clusters, probabilistic clustering methods (e.g. Smyth et al. 1999) are developed, which suggests the possibility (e.g. percentage) that an event could be assigned to each cluster rather than assigning it only to one cluster. If one increases the cluster number, one can expect to decrease such ambiguity in classification. However, the point of clustering analysis is to give a physical insight with minimal groups and not to interpret every single episode.
2.3.5 Self-organizing maps
The input maps themselves determine the degree and types of pattern transitions, hence the “self-organizing” nature of the resulting array. The SOM node array is trained on a sequence of input maps through an artificial neural net technique. The SOM array does not necessarily favor the largest scales in the input data, but rather the scales most relevant to the field for the domain and resolution examined. Consequently, SOMs can extract nonlinear pattern changes in fields, such as shifts in strong gradients. In addition, the pattern at each node is essentially a composite of input maps with similar spatial distribution for the field examined, so that patterns in the SOM array show archetypal patterns of the field examined and directly lend themselves to physical interpretation. Typically, the SOM array displays features having the highest temporal variance in the input data. From this perspective, the SOM array is roughly akin to a transformation of a rotated EOF from spectral space back to physical space.
An advantage of SOMs is that one can identify the nodes where extreme events occur frequently and thus the physical behavior yielding extremes. For example, extreme events may tend to cluster in a small portion of SOM space, thereby allowing identification of LSMPs yielding extreme events. A further advantage is that if more than one group emerges in the SOM-space frequency distribution, then the grouping provides a SOM-determined segregation of different types of extreme events. One can then focus analysis and composites of additional fields (e.g., precipitation, winds, temperature) on only events of the same type. For example, Cassano et al. (2006) used SOMs of sea-level pressure patterns over Alaska to determine which synoptic weather patterns were responsible for extreme wind and temperature events at Barrow, Alaska. They then found robust links between these large-scale synoptic weather patterns and local weather features (precipitation, winds, and temperature). One can construct estimates of the significance of differences in frequency distributions in SOM space through bootstrapping procedures to estimate the likelihood that frequency distributions are not simply the result of random, finite sampling of the pattern space. Thus one can compare frequency distributions between a present and projected climate to assess potential climate changes in LSMPs, or between observational and model climates to assess similarity of observed and simulated LSMPs yielding extremes.
A disadvantage of SOMs is that the array size is pre-determined by the user, and there is no clear, objective guideline for selecting array size. There are, however, some factors that can affect the array-size choice. The issue of significance limits the degree of discretization (number of nodes) one applies to the SOM space. Fine discretization will allow apparent detection of small differences in how different data sets occupy pattern space, but fine discretization will also render very noisy frequency distribution functions of the fields in SOM space, thus undermining detection of any significant differences. Coarse discretization limits the ability of the SOM procedure to resolve features producing the extreme events, so a further disadvantage is that an insufficient array size may obscure grouping that may be present in the data. The training method also requires specification of parameters that govern the training process. A well-trained SOM is insensitive to these choices, but care is needed to ensure such a result. In addition, like some of the other methods described here, the extreme events are defined separately from the SOM analysis.
2.3.6 Machine learning and other advanced techniques
Looking to the future, we note that substantial progress has been made in the field of machine learning for extracting patterns from Big Data. Commercial organizations such as Google and Facebook rely on sophisticated, scalable analytics techniques for mining web-scale datasets. Both supervised and unsupervised machine learning tools could play an important role in extracting spatio-temporal patterns from climate datasets. The technique Deep Belief Networks (Salakhutdinov and Hinton 2012) has been applied with tremendous success to classifying objects in digital images (Krizhevsky et al. 2012) and speech recognition (Hinton et al. 2012). These methods have substantially outperformed existing techniques in the field with the same underlying learning algorithm. While these techniques have not yet been adapted for a multivariate spatio-temporal dataset (such as in climate), research efforts are currently underway to evaluate the performance of such methods in extracting patterns as well as anomalies from datasets. It is too early to discern pros and cons fully for such methods.
2.4 Including large scale patterns in extreme statistics
Application of covariates in extreme value methods (termed “conditional extreme value analysis”) is relatively new to the climate science community, although it has been available to the larger statistics community for some time (Coles 2001). The basic idea of conditional extreme value analysis is to allow the extremal distribution to be dynamic; that is, shifting depending on the observed value of an index of a climate mode or LSMP (the index would be an example of a “covariate”). The book by Coles includes an example in which annual maximum sea level is related to a climate mode, the Southern Oscillation. In their study of changes in extreme daily temperatures, Brown et al. (2008) used the NAO as a covariate in addition to a trend component.
Such techniques have proven useful in connecting extreme temperatures to LSMPs. Sillmann and Croci-Maspoli (2009) and Sillmann et al. (2011) used a blocking index as a covariate for extremely cold European winter temperatures and found that extreme value distributions (based on block minima) were better fit and long period return values were somewhat colder. Furthermore, they concluded that projected future extremely cold events in Europe were less influenced by atmospheric blocking because of projected shifts in North Atlantic blocking patterns. Photiadou et al. (2014) used a similar technique (but based on the POT approach rather than block maxima) to connect blocking and other indices to European high temperature events finding that while El Nino/Southern Oscillation (ENSO) does not exert much influence on extremely high temperature magnitudes or duration, the North Atlantic Oscillation (NAO) and atmospheric blocking do. However to date, such covariate techniques relating atmospheric blocking to extreme temperatures have not been applied to North America. Many physically based covariate quantities potentially offer insight into the mechanisms behind extreme temperature events and the response to changes in the average climate. Good North American candidates for covariates include indices measuring modes of natural variability such as those describing the ENSO, the Pacific Decadal Oscillation (PDO), the NAO, the North Atlantic Subtropical High (NASH), and various blocking indices.
The ETCCDI indices were designed for climate change detection and attribution purposes rather than for exploring the mechanisms causing extreme events. They are not ideal for connecting extreme temperature events to LSMPs and they are not descriptive of particularly rare events. However, ETCCDI indices are designed to be robust over the observational record and have been calculated and described for CMIP5 models by Sillmann et al. (2013a) (see Table 1). ETCCDI indices are intended to be applied globally and be meaningful in areas of sparse observations. The relatively dense network of North American observations since the beginning of the twentieth century permits the construction of more specialized LSMP extreme indices linked to specific extreme events. Grotjahn (2011) defines an index that is an unnormalized projection of key parts of a target ensemble LSMP onto a daily map of the corresponding variable. (He combined such projections onto 850 hPa temperature and 700 hPa meridional wind to form his ‘circulation index’.) His target ensemble members are from dates satisfying criteria listed in Table 3. The key parts of the fields used are those where all the extreme events in the training period were consistent, at least in having the same anomaly sign. Grotjahn (2011) found that extreme values of such an index (based on upper air data) occurred on many of the same dates as extreme values of surface stations in the California Central Valley (CCV). Statistically significant relationships exist between extreme values of this circulation index and both the rate of CCV daily maximum temperature exceeding a high threshold and the distribution of the excess over the threshold (Katz and Grotjahn 2014). Grotjahn (2013) used such an index to show that a particular climate model was notably under-predicting the occurrence frequency (by half) of CCV hot spells. Grotjahn (2015) used such an index to show how that same model compared with a 55 year historical record and what the model implied for CCV hot spells during the last half of the twentyfirst century under two representative concentration pathways (RCPs) of greenhouse gases.
3 Large scale meteorological patterns related to extreme temperature events
Intraseasonal extreme temperature events (ETEs) are almost always associated with regional air mass excursions induced by circulation anomalies that are part of large-scale meteorological patterns (LSMPs). LSMPs can include synoptic features (e.g., midlatitude cyclones; Konrad 1996) that enhance the ETE and often with scales similar to a teleconnection pattern (though the nodes may not align and ETE onset, itself, may impact the teleconnection pattern, Cellitti et al. 2006). In some cases, the LSMP is interpreted as a juxtaposition of teleconnection patterns that leads to ETE events (Lim and Kim 2013). Heuristically, the role of LSMPs in producing ETEs could be considered the result of either (1) a direct contribution to the large-scale circulation that facilitates the air mass excursion or (2) the indirect modulation of sub-scale variability, such as regional modulation of storm track behavior by blocking patterns. Besides such dynamically driven impacts, there exist possible local impacts related to the interaction of the LSMP with local topography or coastline features, leading to possible local symmetries in the response pattern (e.g., Loikith and Broccoli 2012). Current knowledge of the remote forcing, dynamics and local forcing of LSMPs associated with ETEs is summarized next.
3.1 Remote forcing of LSMPs and ETEs
3.1.1 Connection to low frequency modes of climate variability
There have been pronounced episodes of climate modes influencing ETEs during recent winters. Cold extremes over Europe and the southeastern United States during recent winters (2009–2010 and 2010–2011) were primarily accounted for by the anomalous blocking associated with persistent episodes of large amplitude negative phase of the NAO (Guirguis et al. 2011). There is also evidence of an important role for stationary Rossby wave patterns in contributing to North American temperature extremes during summer (Schubert et al. 2011; Wu et al. 2012). These wave patterns appear to arise from internal forcing associated with intraseasonal transient eddies (Schubert et al. 2011).
As discussed above, two more commonly recognized remote influences upon North American ETEs are associated with ENSO and the PDO (e.g., Westby et al. 2013), both involving local sea surface temperature anomalies and atmosphere–ocean coupling. These generally operate in conjunction with PNA-like teleconnection patterns that extend from the coupling region downstream into North America. Similar to the effect of climate modes, the impact of remote forcing upon warm season ETEs is partly limited by the relative inactivity and spatial extent of climate modes, which serve as horizontal pathways for Rossby wave energy between the remote forcing region and the local surface response (Schubert et al. 2011).
3.1.2 Connection to sea ice and snow cover
The atmospheric response to the Arctic sea ice reduction is thought to be Arctic warming and destabilization of the lower troposphere, increased cloudiness, and weakening of the poleward thickness gradient and polar jet stream (Francis et al. 2009; Outten and Esau 2012). As the Arctic warms faster than lower latitudes (so-called Arctic Amplification), the meridional temperature gradient at higher latitudes is likely to weaken altering the polar jet stream according to thermal wind balance. Changes in the high latitude jet stream in turn have the potential to impact weather conditions at middle and high latitudes. For example, during winter an enhanced westerly jet over the North Atlantic can help maintain relatively mild conditions over northwest Europe via heat transport from the Atlantic. Cohen et al. (2014) review three “pathways” by which Arctic amplification may impact extreme weather events in mid-latitudes. In principle, Arctic amplification may lead to regional alterations in the structure of storm tracks, jet streams and planetary waves. In recent years, considerable attention has been focused on the role of Arctic amplification-induced changes to the jet stream (Francis and Vavrus 2012; Liu et al. 2012; Barnes 2013; Screen and Simmonds 2013). Francis and Vavrus (2012) found that a weaker zonal flow (i.e., polar jet) from weakened meridional temperature gradient slows the eastward Rossby wave progression and tends to create larger meridional excursions of height contours and associated temperature displacements resulting in a higher probability of extreme weather. In a similar vein, Liu et al. (2012) argue that the circulation change due to the decline of Arctic sea ice leads to more frequent events of atmospheric blocking that cause severe cold surges over large parts of northern continents. Francis et al. (2009), Overland and Wang (2010), Jaiser et al. (2012) and Lim et al. (2012) found that there is a delayed atmospheric response to the Arctic sea ice. Specifically, the Arctic sea ice extent in summer to fall influences the atmospheric circulation in the following winter over the northern mid- to high-latitudes, affecting the seasonal winter temperature and subseasonal warm/cold spells. Evidence presented in more recent studies (Barnes 2013; Screen and Simmonds 2013), however, suggests that the role of the mechanism put forth by Francis and Vavrus (2012) is uncertain at best. More generally, Cohen et al. (2014) conclude that our understanding of the mechanistic link between ongoing Arctic amplification and mid-latitude extreme weather is currently limited by shortcomings in relevant data records, physical models and dynamical understanding, itself. As such, the likely future impact of Arctic amplification upon extreme weather is highly uncertain.
Arctic amplification is also linked to long-term variability in high latitude snow cover. An analogous linkage between autumnal variability in Eurasian snow cover and wintertime ETE events over North America has been noted (Cohen and Jones 2011). In this case, autumnal snow cover anomalies induce a subsequent weakening of the stratospheric polar vortex during winter which, in turn, leads to a persistent negative phase episode of the tropospheric AO favoring North American cold events (Cohen et al. 2007). In addition, there is evidence that the changing Arctic sea ice extent may be linked to changes in the autumnal advance in Eurasian snow cover (Cohen et al. 2013). As for Arctic amplification, however, there is considerable uncertainty regarding the statistical robustness and physical nature of the Eurasian snow cover influence upon mid-latitude extreme weather (Peings et al. 2013; Cohen et al. 2014).
3.1.3 Large-scale climate “markers” for climate model assessment
Representation of fundamental climate modes One obvious essential minimum requirement for climate models to properly represent the modulation of ETEs by climate modes is the extent to which the models are able to represent the primary climate modes, themselves. Thus, fundamental markers for model assessment are metrics that measure the representation of key extratropical climate modes including those internally forced on intraseasonal time scales (PNA, AO/NAO and atmospheric blocking) and those externally forced on longer time scales (the extratropical response to ENSO and the atmospheric part of the PDO). Atmospheric models have had historical difficulty in representing some types of intraseasonal low frequency variability (Black and Evans 1998). A particular problem is an under-representation of atmospheric blocking activity (Scaife et al. 2010). In a similar vein, the representation of externally forced extratropical modes connected to ENSO and PDO depends on how well the coupled climate models simulate the associated oceanic phenomenological behavior.
Correlation coefficients between the loading pattern of observation and that of each CMIP5 model for NAO-like and PNA-like modes
The tropical large-scale circulation response to tropical diabatic heating
Upper tropospheric meridional vorticity gradients in the subtropics
Barotropic deformation structures in the jet exit regions
The structure and intraseasonal variability of the Atlantic and Pacific storm tracks
The authors recommend that model validation activities concentrate upon the above features in order to uncover likely sources for existing model deficiencies in leading teleconnection patterns.
3.2 Dynamics of LSMPs
3.2.1 Diagnostic tools to study dynamics of LSMP onset/decay
The onset and decay of LSMP structures linked to ETEs generally occur on relatively short time scales. As such, the range of dynamical forcing mechanisms that can directly account for LSMP time evolution is limited to internal atmospheric processes (given the relatively long time scales associated with boundary forcing). The atmospheric processes affecting LSMP evolution can be local or remote. For example, local synoptic-scale cyclogenesis can usher in a regional Arctic air outbreak while a transient episode of the PNA teleconnection pattern can provide a remote downstream influence on North American extremes. In some cases it is an optimal juxtaposition of local and remote influences operating on multiple time scales that is required to produce an extreme event (e.g., Dole et al. 2014).
Most ETE events are associated with lateral air mass excursions, which are induced by large-scale circulation anomalies in the lower troposphere forming the LSMP (e.g., Loikith and Broccoli 2012). The anomalous circulation serves as a “dynamical trigger” for ETE events. The delineation of LSMP dynamics is treated as a two-stage process: First it is of interest to assess whether the main energy source(s) are local or remote. Once the energy source location is determined, the second stage is to assess the specific physical mechanism providing the proximate energy source in this location. An effective means for assessing the source location regions for large-scale atmospheric waves is the application of wave activity flux analyses (Plumb 1985; Takaya and Nakamura 2001). The wave activity flux is parallel to Rossby wave group velocity and traces out a three-dimensional pathway between a wave source region (of flux divergence) and a wave sink region (of flux convergence).
Possible primary physical mechanisms providing wave activity sources in this context include large-scale barotropic growth, baroclinic growth/instability and nonlinear forcing by synoptic-scale eddies (Evans and Black 2003). These mechanisms may be augmented via secondary feedbacks related to internal diabatic processes or interactions of the LSMP with the local topography or land surface. Past studies have introduced comprehensive dynamical frameworks for studying the dynamical mechanisms leading to the growth and decay of large-scale circulation anomalies (e.g., Feldstein 2002, 2003). These are based upon a local analysis of tendencies in either the geopotential (Evans and Black 2003) or streamfunction (Feldstein 2002) fields. In these studies local tendencies are decomposed into separate forcing terms that can be related to distinct physical processes. The two-stage process outlined above is a generally useful means for the dynamical diagnosis of LSMP life cycles in both observations and climate model simulations.
3.2.2 North American Arctic air mass formation
Both dynamics and thermodynamics have roles in the formation of extreme cold-air masses. Wexler (1936, 1937) postulated that the cold air was formed at the surface by radiative cooling from a snow-covered ground under clear, windless conditions, creating an intense temperature inversion restricted to a very shallow layer at the surface. Finding that the Wexler model could not adequately explain the depth of the cold layer observed in soundings, Gotaas and Benson (1965) showed that the existence of suspended ice crystals were crucial to upper-level cooling. Curry (1983) later modeled the effect with the introduction of condensate, particularly ice-crystals, in the layer. In an experiment without moisture, the inversion that formed after 2 weeks of radiative cooling was less than 1000 meters deep. Only with the addition of ice crystals did the inversion rise above 900 hPa. More recently, Emanuel (2009) found that the rates and depth of cooling in his model were sensitive to the amount of water vapor and clouds present.
Turner and Gyakum (2011), in their composite study of 93 Arctic air mass formations in Northwest Canada, found that the cold air mass lifecycle is a multi-stage set of processes. During the first stage, snow falls into a layer of unsaturated air in the lee of the Rockies, causing moisture increases in the sub-cloud layer. Simultaneously, the mid-troposphere is cooled by cloud-top radiation. On the second day, snowfall abates, the air column dries and clear-sky surface radiational cooling prevails, augmented by the high emissivity of fresh snow cover. The surface temperature falls very rapidly, as quickly as 18 °C day−1. On the third day, after near-surface temperatures fall below their frost point, ice crystals and, nearer the surface, ice fog form. At the end of formation, there is cold-air damming (Forbes et al. 1987; Fritsch et al. 1992), with a cold pool and anticyclone in the lee of the Rockies, lower pressure in the Gulf of Alaska and an intense baroclinic zone oriented northwest to southeast along the mountains.
3.2.3 CONUS wintertime cold air outbreaks
West coast events (e.g., CA Central Valley events) Compared with cold air outbreaks affecting the eastern and central parts of the US, less has been written about outbreaks affecting the western US. Extreme cold over California is related to a large scale pattern that brings cold air from the Arctic and northern Canada without crossing the Pacific, and hence over the Rocky Mountains (Grotjahn and Faure 2008). As discussed in Grotjahn and Faure, the large scale pattern is similar to outbreaks affecting the areas east of the Rockies however with a primary difference being a small but statistically significant ridge over the southeastern US (see Fig. 2a, b). This ridge precedes the CAO by several days, including the large ridge that develops subsequently over Alaska. A cold air mass between these ridges is directed further westward (a more northerly or northeasterly flow of the cold air) than for eastern CAOs. The distortion of the flow by that southeastern ridge enables the cold air mass to cross the Rockies. Loikith and Broccoli (2012) capture a portion of this pattern in geopotential heights at 500 hPa using their ‘grand composite’ technique. More local to the area of the extreme, colder temperatures are associated with an adjacent surface high (e.g. the conceptual model of Colle and Mass 1995). Favre and Gershunov (2006) pursue this link to anticyclones affecting western North America; and develop indices based on frequency and central pressure of transient cyclones and anticyclones in the eastern North Pacific. Their ‘CA’ index is the difference, cyclone minus anticyclone, in strength and frequency. Negative CA values are correlated with the coldest 10 % of winter minimum temperatures.
Eastern US events (i.e., east of Rockies) Several statistical and synoptic studies either directly or indirectly relate to the topic of US CAOs occurring east of the Rockies (e.g., Konrad and Colucci 1989; Colle and Mass 1995; Konrad 1996; Walsh et al. 2001; Cellitti et al. 2006; Portis et al. 2006). Common precursors to eastern CAO events include anomalously high surface air pressure over western Canada (linked to a polar air mass), occurrences of the negative (positive) phase of the NAO (PNA) teleconnection pattern and/or an anomalously weak stratospheric polar vortex. CAO onset is characterized by the south/southeastward advection of the polar air mass in association with one or more of the following synoptic LSMP features: southward extension or propagation of the surface high pressure from Canada, surface cyclogenesis over the eastern US, anomalously low 500 hPa geopotential heights over the Great Lake region and the southeastward movement of an upper level shortwave from Canada.
Distinguishing between east coast (EC) and Midwest (MW; just east of the Rockies) CAO events, EC events result from the geostrophic cold advection enhanced by southeastward propagation of the surface high pressure system (with winds often amplified by surface cyclogenesis further east) leading to detachment of the polar air mass from its source (Konrad and Colucci 1989; Walsh et al. 2001). MW events are linked to the southeastward meridional extension of surface high pressure from Canada leading to unusually lengthy cold air transport (Walsh et al. 2001). For MW events occurring just east of the Rockies, the cold advection is related to low-level northerly ageostrophic flow within a region of cold-air damming against the topography (Colle and Mass 1995). In all cases, the salient LSMP features are synoptic-scale in nature and, thus, governed by generally well-understood dynamical physical principles considered to be well-represented by weather and climate models. The CAO process, itself, is largely due to radiative cooling (during polar air mass formation and early stages of CAO onset) and southward air mass transport (by the horizontal wind) partly offset by adiabatic warming induced by large-scale subsidence (Konrad and Colucci 1989; Walsh et al. 2001; Portis et al. 2006). Although the proximate physics associated with the CAO process itself is reasonably well characterized, our knowledge of the physics of the LSMP patterns is implicit and depends on subjective associations (between advective circulation features and synoptic LSMPs) that have been made in earlier studies. As such, there is an existing need to (a) more objectively link critical advective CAO circulation features to dynamical entities (e.g., potential vorticity) and (b) subsequently assess the physical origin of these dynamical entities.
3.2.4 Summer heat wave events
Hot spells over North America have an intense upper level ridge as one expects hydrostatically. This ridge is seen in Figs. 2 and 6. There can also be a shallow layer of relatively lower pressure, sometimes called a ‘thermal low’. Hot spells occur from a combination of factors: a hot air mass displaced from its normal location (displaced from the southwestern desert), strong subsidence (causing adiabatic warming), and solar heating. Solar heating is more effective when surface latent heat fluxes are constrained (by drought, for example). The first two factors are enhanced and help define the LSMP for a hot spell; there may be an upstream trough (e.g. Figs. 2, 6) that enhances southwesterly advection of warmer air. Other factors influence how the LSMP develops. What follows are details for different regions of North America.
Gershunov et al. (2009, hereafter GCI) make the novel distinction of considering separately heat waves that have high daytime maximum and events with high nighttime minimums. In so doing they uncover an LSMP with unusually high values of precipitable water (PW) over their region of interest occurring during their ‘nighttime’ heat waves (but not their ‘daytime’ events). GCI emphasize a trend of increasing occurrence of ‘nighttime’ events. Gershunov and Guirguis (2012) also find that trend in all their sub-regions of California and also show another trend: increasing longitudinal extent of Central Valley heat waves. GCI show maps of anomalous geopotential heights which also show the west coast ridge (height anomaly centered over Washington State) with upstream (and downstream) troughs, consistent with Grotjahn and Faure. GCI show the LSMP for SLP (high over the Great Plains and low off the west coast) and remark that both daytime and nighttime events have a corresponding general southeasterly flow of air out of the desert southwest. This flow occurs throughout the lower troposphere with associated high frequency (<7 day period) heat fluxes that are prominent just off the coast (Grotjahn 2015); and consistent with setting up the offshore pressure gradient (to oppose a cooling sea breeze).
For heat waves affecting the western areas of Washington and Oregon, Bumbaco et al. (2013, hereafter BDB) show an upper level (500 hPa) ridge over the west coast and trough upstream over the Gulf of Alaska. That trough and ridge pattern is generally similar to the pattern for California heat waves, including the tropospheric height anomaly centered at the coast (Grotjahn 2011). Using a regression approach Lau and Nath (2012) find a ridge and trough further upstream (over the western North Pacific) that are stronger several days prior to the event onset; both the location and timing of the ridge and trough are similar to those shown in Grotjahn and Faure. Similar to Grotjahn (2011, 2013, 2015), BDB find the lower tropospheric temperature anomaly (850 hPa) to be centered at the west coast of North America. Hence, BDB find negative values of sea level pressure anomaly centered offshore that sets up an offshore low level pressure gradient and offshore (northeasterly) flow (again similar to California heat waves). Consistent with the heat fluxes shown in Grotjahn (2015), the near-coast thermal trough appears to migrate from the southwestern deserts across the length of California to finally reach western Oregon and Washington (Brewer et al. 2012).
Midwest events Heat waves affecting the Midwest have anticyclonic flow at mid-levels, either as a closed anticyclone or as a strong ridge (Klein 1952; Karl and Quayle 1981; Namias 1982; Chang and Wallace 1987; Lyon and Dole 1995; Kunkel et al. 1996; Livezey and Tinker 1996; Palecki et al. 2001; Meehl and Tebaldi 2004; Lau and Nath 2012; Loikith and Broccoli 2012; Teng et al. 2013), with associated clear skies allowing maximum solar heating of the surface as well as adiabatic warming from subsidence.
For at least some events, the continental anticyclonic flow at mid-levels is part of a larger pattern of anomalies with remote centers over the Pacific and Atlantic (Namias 1982; Chang and Wallace 1987; Livezey and Tinker 1996; Lau and Nath 2012; Loikith and Broccoli 2012; Teng et al. 2013), with some suggestion that the anomalies are generated in the Pacific (Namias 1982; Lyon and Dole 1995; Livezey and Tinker 1996; Teng et al. 2013), with possible predictability (Teng et al. 2013). Teng et al. (2013) have noted the similarity of the circulation pattern to a Rossby wave number 5 wave train in the jet stream waveguide. The commonality of the wave train pattern structure and evolution among these studies is difficult to assess, though.
Studies of the surface flow have been less common, but negative SLP anomalies have been noted (Chang and Wallace 1987; Lau and Nath 2012; Loikith and Broccoli 2012), and Lau and Nath (2012) have shown anomalous southerly (or southwesterly) flow at the surface, consistent with strong horizontal warm advection. The regional-scale flow also interacts with local mechanisms, particularly the Urban Heat Island (UHI) effect (Kunkel et al. 1996; Palecki et al. 2001). Another common element identified in several studies is the presence of drought (Chang and Wallace 1987; Karl and Quayle 1981; Namias 1982; Lyon and Dole 1995) as well as simultaneous precipitation deficits (Lau and Nath 2012).
Eastern events While the large-scale meteorological patterns associated with heat waves in the eastern US have not yet seen much study, they have been examined for the Northeast and Central US Gulf Coast in Lau and Nath (2012) and for piedmont North Carolina by Chen and Konrad (2006). For both the Northeast and Gulf Coast, Lau and Nath found a mid-level ridge centered over the region as part of a wave train and precipitation deficits, broadly similar to the results for the Midwestern regions. For piedmont North Carolina, Chen and Konrad showed that, at upper levels, a strong ridge over or just upstream of the region was a common feature along with, at lower levels, adiabatic warming associated with descent from the Blue Ridge Mountains into the piedmont region.
4 Modeling of temperature extremes and associated circulations
Temperature extremes have some predictability at weather and climate time scales that can be exploited by models for short term predictions and long term projections. At weather time scales the predictability of North American extreme temperature events (ETEs) is largely dependent upon the nature of the LSMPs that help organize their occurrence. The greatest predictability is expected to occur during the boreal cool season when ETEs are, at least in part, influenced by low frequency modes (PNA, NAO and blocking events) (Westby et al. 2013) with intrinsic time scales of several days to weeks (Feldstein 2000). Loikith and Broccoli (2012) illustrated that the local LSMPs linked to ETEs generally exhibit synoptic spatial scales rather than planetary scales, though a wave train can be long (e.g. Grotjahn and Faure, 2008; Fig. 2). Given the essential role of synoptic-scale disturbances such as east coast cyclones and southward moving polar anticyclones over the Midwest (Konrad and Colucci 1989; Walsh et al. 2001) in ETEs, pointwise predictability of ETEs is ultimately limited by our ability to forecast the details of synoptic-scale phenomena several days in advance (Hohenegger and Schär 2007).
Hence a significant challenge for models to predict temperature extremes is their ability to predict or simulate synoptic scale phenomena, low frequency modes that provide a large-scale meteorological context, and small scale atmospheric processes and land surface processes that influence the surface heat fluxes. Both dynamical and statistical models have been developed and used to simulate and project changes in temperature extremes. This section briefly summarizes the methods and skills of the models, while analysis of observed trends and projections of future trends are summarized in Sect. 5.
4.1 Global and regional climate model skill in simulating temperature extremes
To get a sense of model errors relative to the observational uncertainty, the 3 rightmost columns of Fig. 8, show the error of three other reanalyses considered in Sillmann et al. (2013a) relative to the model median RMSE. The two best models (leftmost in Fig. 8) are competitive with these other reanalyses but most models’ RMSE relative to ERA Interim are significantly larger than the disagreement among reanalyses for many fields. Reanalysis disagreements in the cold night metrics (tnn, tn10p) and a cold day metric (tx10p) are notable and a larger number of models perform better in these fields relative to this crude measure of observational uncertainty.
Sillmann et al. (2013a) found that the spread amongst CMIP5 temperature extreme indices tends to be smaller than that of the CMIP3 models (earlier version of the CMIP5), indicating reduced uncertainty. They identified an increase in the monthly maximum daily maximum temperature and a decrease in the monthly minimum daily minimum temperature by the CMIP5 models over the high northern latitudes, compared to CMIP3. Analysis of 20 year return values in the CMIP3 models by Zwiers et al. (2011) found that models tend to warm the minimum temperature less than observed and that they warm the maximum temperature more than observed over the latter half of the twentieth century.
Westby et al. (2013) analyzed reanalysis data and CMIP5 historical simulations to study the statistics and low frequency mode modulation of wintertime CAOs and warm waves (WWs) occurring over the continental United States during 1949–2011. The observational results indicate (a) a lack of significant long-term trends in ETE frequency and (b) a seasonal modulation of CAOs (WWs) by the NAO and PNA (NAO, PNA, ENSO and PDO) patterns (Fig. 4). Similar behavior is found in the CMIP5 models, showing CAOs are mainly modulated by NAO in the southeastern US and by PNA in the northwestern US, while WW frequency is modulated by NAO over the eastern US and by a combination of PNA, PDO and ENSO over the southern US. The study also found modest influence of ENSO on WWs over the southern US, in good agreement with Lim and Schubert (2011). Comparison of WWs and CAOs between the CMIP5 models and observation indicated that the models tend to overestimate WW frequency, but underestimate CAOs frequency, although using a larger ensemble of CMIP5 models, Gao et al. (2015) found a better agreement between CAO frequency from CMIP5 models and reanalysis datasets.
Heat waves are associated with anomalous large-scale circulation patterns well resolved by GCMs. Grotjahn (2013, 2015) describe how the LSMPs for California hot spells simulated by a CMIP5 climate model compare with the observed LSMPs; while the model generates a similar large scale pattern and included the needed large scale sinking, the model is unable to simulate sea breezes so the effect of LSMP blocking of sea breezes during extreme heat episodes is likely not captured by the global model. In general, heat wave intensity and duration can vary at local to regional scales because surface temperature is influenced by processes such as land surface fluxes, turbulence and winds, and clouds and radiation. For example, land–atmosphere interactions were found to play an important role in European summer heat waves of the past decades (Fischer et al. 2007). Using a regional model, Jaeger and Seneviratne (2011) and Lorenz et al. (2010) highlighted the key role of soil moisture memory on heat wave intensity and persistence, respectively. Using observations and a CMIP3 model, Gershunov and Douville (2008) showed a strong negative relationship between preceding winter and spring precipitation and summertime heat in the central, Midwestern, and eastern US. In regions with limited soil moisture, surface temperature can respond more strongly to the large-scale circulation anomaly that leads to clear sky and strong surface solar heating because surface cooling from evapotranspiration is limited. As soil moisture is influenced by antecedent precipitation and surface properties such as land cover, topography, and soil characteristics that vary at local to regional scales, heat wave intensity and duration can vary similarly, within the context of large-scale influence by a circulation anomaly. Thus the ability of climate models to simulate and predict heat waves is also dependent on physics parameterizations and grid resolution.
Lau and Nath (2012) analyzed GCM simulations performed at 200 km and 50 km grid resolutions and found considerable resemblance between the temperature anomaly patterns from the coarse and fine resolution simulations, except for some local details generated by the higher resolution model. Both models were able to capture realistic heat wave intensity, duration, and frequency in various key regions of North America and the synoptic features accompanying the warm episodes as revealed in the North American Regional Reanalysis. Kunkel et al. (2010) compared regional climate simulations performed at 30 km resolution over North America driven by two GCM historical simulations with the respective GCM simulations and observations. Although the two GCM simulations have opposite temperature biases, downscaling by the RCM provided improved simulations of heat waves in both cases, with the overall averaged biases reduced by a factor of two compared to the GCMs. Gao et al. (2012) compared a regional simulation at 4 km grid resolution over the eastern US driven by a GCM with observations and the GCM simulation. They found statistically significant improvements in heat wave intensity and duration in the regional simulation compared to the global simulation in 14 out of 16 states in the eastern US and attributed the improvements to high-resolution topography and land use information that impact local surface temperature.
In summary, global climate models have demonstrated useful skill in simulating heat waves and cold air outbreaks over North America and their linkages to low frequency variability such as NAO, PNA, and ENSO. However, models generally underestimate the combined influence of low frequency modes on temperature extremes, with PDO and its modulations on temperature extremes being most notably deficient. Although heat waves and their LSMP are well resolved by global models, some studies have demonstrated improved skill in simulating heat waves using regional models that may better capture regional processes such as land–atmosphere interactions influenced by surface heterogeneity.
4.2 Statistical modeling of temperature extremes
Most statistical methods for downscaling of temperature can be thought of as extensions of the model output statistics and related approaches to statistical weather forecasting (e.g., Wilks 2006). The most popular method is multiple regression analysis (e.g., Easterling 1999) but it cannot be expected to represent extremes well, as only the conditional mean of temperature given the values of the predictor variables (such as indices of LSMPs) is statistically modeled. In particular, the error term in the regression equation is usually assumed normally distributed, which is not necessarily realistic for temperature extremes.
Recently, bias correction techniques involving quantile mapping (Thrasher et al. 2012), or quantile matching (Ho et al. 2012), have become a popular way to adjust the entire probability distribution of temperature. Because of the mismatch in spatial (and sometimes temporal) scales involved in downscaling, the downscaled temperatures typically do not possess high enough variance. This issue is crucial for extremes, with it being preferable to increase the variance through randomization rather than direct “inflation” (Maraun 2013; von Storch 1999). Ideally, quantile mapping for the extreme tails of the temperature distribution should use statistical methods based on extreme value theory (e.g., Coles 2001). Kallache et al. (2011) proposed such a method, in which the quantile mapping involves fitting the Generalized Pareto distribution (GPD) to the upper (or lower) tail of the distribution. Although their application was to precipitation extremes, the technique should apply equally well to temperature extremes.
The statistical methods for relating extreme temperatures to LSMPs and climate modes, described earlier in Sect. 2, could also be applied in the context of statistical downscaling. In this approach, the extremal distribution (GEV or GPD) is conditioned on indices of LSMPs or climate modes, as well as possibly on other covariates in the context of statistical downscaling. Brown et al. (2008) relate daily maximum and minimum temperatures to the NAO index; Sillmann et al. (2011) relate European CAOs to a blocking index; Katz and Grotjahn (2014) relate California hot spells to an associated LSMP index. The chief limitation is that, given the effective reduction in data when only considering extremes, not as many predictors necessarily could be included as in more conventional statistical downscaling.
5 Observed and projected trends of temperature extremes
5.1 Observed trends and variability
Traditional methods in climate research for detecting trends in temperature extremes tend to focus on the relative frequency of exceeding a high threshold or of falling below a low threshold, as well as indices derived from such rates (Karl et al. 1996; Gleason et al. 2008). Such approaches generally fail to give a complete picture of trends in temperature extremes as they neglect the intensity of the event. It should be noted that other forms of indices are also commonly used to monitor changes in temperature extremes (Zhang et al. 2011).
Extreme value statistics (Coles 2001) formally account for both the rate of exceeding a high threshold and in the excess over a high threshold so they can be applied to detect trends in each of these two components (Katz 2010).A fundamental result in extreme value theory is that block maxima (e.g., highest daily maximum temperature each summer), suitably normalized, have an approximate generalized extreme value (GEV) distribution. In practice, it is more efficient and can be more informative to fit the GEV indirectly using the peaks over threshold (POT) approach, which involves a Poisson distribution to approximate the rate of exceeding a high threshold; and a generalized Pareto (GP) distribution to approximate the excess over the high threshold. Possible trends in extremes can be introduced into these extreme value distributions with time as a covariate. That is, the location, scale, and shape parameters of the GEV (or, equivalently, the rate parameter of the Poisson distribution and the scale and shape parameters of the GP distribution) can shift with time (Coles 2001). See Brown et al. (2008) for examples of the application of the POT approach to trend analysis of temperature extremes.
Peterson et al. (2013) applied the POT approach with a threshold corresponding to the 99th percentile for the upper tail (or 1st percentile for the lower tail) specific to each grid point in the US. To avoid temporal dependence of extremes, only the single highest (or lowest) daily temperature within a run of consecutive days exceeding the threshold was retained (termed “runs declustering”). Using the equivalent parameterization in terms of the GEV distribution, a linear trend in the location parameter of the GEV was fitted by the technique of maximum likelihood (Coles 2001), assuming that the scale and shape parameters were constant over time. To adjust for seasonality, the analysis was performed in terms of daily anomalies.
Figure 1 shows the estimated change in 20-yr return levels between 1950 and 2012, along with an indication of statistical significance. The broadest region of warming occurs for the cold tail of minimum temperature (panel d). Cooling occurs in the upper tail of both daily maximum (panel a) and minimum (panel c) temperature in the southeastern US. Although these results are not explicitly in terms of heat waves or cold air outbreaks, they do reflect changes in extremes that are often part of such multi-day events (Furrer et al. 2010).
Atmospheric reanalyses have provided a significant source of data for weather and climate research because they assimilate multitudes of available observing systems and provide continuous fields for regions or variables with sparse or no direct observations. Utilizing reanalyses in the calculation of extreme indices should be approached with a strong validation strategy (as discussed in Zhang et al. 2011). In the case of temperature, processes near the surface can influence the distribution of temperature. However, Simmons et al. (2010) show that ERA Interim reanalysis surface temperature reproduces that of HadCRUTv very closely, owing to the correction of soil water by the analysis of near-surface atmospheric water and temperature. Further, Bosilovich (2013) shows that even without surface analysis, reanalyses can provide robust interannual variability of seasonal surface temperature at regional scales. Furthermore, reanalysis temperature, augmented with observations, can provide a substantially improved representation of the diurnal cycle of temperature (maximum and minimum) consistently over the global land mass (Wang and Zeng 2013).
Walsh et al. (2001) used global reanalysis data to study cold air outbreaks and identified Midwest, Gulf Coast, and East Coast as regions influenced by CAOs. They did not identify any apparent trend in CAOs in North America between 1948 and 1999, although analysis from a single station suggests that extreme outbreaks may have been more frequent in the late 1800s and early 1900s. Extending the study to include the recent decade from the global reanalysis data may provide further confidence in potential trends in CAOs in North America.
The twentieth century reanalysis (Compo et al. 2006, 2011) provides objectively analyzed four-dimensional data on a 2° global grid. With assimilation of synoptic surface pressure available since the late nineteenth century, this dataset covers the period from 1871 to 2012 and offers a unique opportunity for analysis of long-term trend in temperature extremes and the associated LSMP. Compo et al. (2011) showed encouraging comparison of this surface-based reanalysis product with other reanalyses that make use of upper-air and satellite data. Future work utilizing this dataset could potentially further advance understanding of variability and trends in LSMPs that influence temperature extremes.
5.2 Projected trends
The CMIP3 and CMIP5 multimodel ensembles have been used to investigate projected changes of temperature extremes for the next couple of decades (e.g., mid-twentyfirst century) (e.g., Tebaldi et al. 2006; Orlowsky and Seneviratne 2012; Sillmann et al. 2013b; Wuebbles et al. 2014). Climate change simulations in the CMIP5 multimodel ensembles showed greater changes in ETCCDI based on daily minimum temperatures than in ETCCDI based on daily maximum temperatures. Also, the strong emission scenario such as representative concentration pathways (RCPs) 8.5 (Moss et al. 2010; van Vuuren et al. 2011) produced more definitive changes in temperature extreme statistics than weaker emission scenarios (e.g., RCP 2.6 and 4.5). Furthermore, the changes in ETCCDI under RCP8.5 tend to be greater than changes under any of the scenarios in the special report on emission scenarios (SRES) (Nakicenovic et al. 2000) used in CMIP3. The spatial distributions of Sillmann et al. (2013b) showed that the northern high latitude regions have the strongest increase in the minimum daily minimum temperature, while changes in the maximum daily maximum temperatures tend to be evenly distributed globally. The study also noted that the percentile indices such as warm and cold nights, exhibit the highest increases in the tropical regions. This is a result of differences in interannual temperature variability between the two regions, which is relatively large in mid- to high latitudes but small in the low-latitudes.
Seasonality in the projected changes of these two indices is a reflection of the complex mechanisms affecting changes in extreme and mean temperatures in North America. Changes in snow cover certainly affect the future winter mean and extreme temperatures. Likewise, decreases in soil moisture affect the future summer mean and extreme temperatures. The additional possibility of biases and changes in LSMPs affecting extreme temperatures must be considered in light of these and other mechanisms. For example, van Oldenborgh et al. (2009) found that changes in the large-scale circulation, including a shift towards a more westerly circulation and the North Atlantic current need to be better simulated especially in winter and spring for more realistic simulations of warming over Western Europe in recent decades. Gershunov and Guirguis (2012) noted that three out of four GCMs analyzed did not adequately capture the synoptic causes of California heat waves. When sub-regional heat waves are defined relative to the changing local climate (a ‘non-stationary’ perspective), coastal heat waves have greater impact than those inland even though the summertime average warming is stronger inland. Sillmann and Croci-Maspoli (2009) found that European blocking events influence particularly the winter cold extremes in Europe. For future projection, decrease in blocking frequency with increase in maximum blocking duration was simulated under the A1B scenario. Also, the blocking pattern shifts northeastward, affecting a larger part of Europe by giving rise to anomalously cold winter months.
To partly address the question of whether changes in heat waves in the future are associated with a shift in the daily maximum temperature or changes in LSMP, Lau and Nath (2012) noted that the probability distribution functions (PDFs) for the current and future climate over most regions that experience heat waves in the US have similar shape but were shifted by the mean warming in the daily maximum temperature, and the magnitude of skewness of the PDFs exhibited only minor changes from the current to future climate. Further statistical manipulations of the model outputs suggest that the increase in heat wave intensity and frequency in the future is primarily associated with a shift in the daily maximum surface temperature, which suggests that changes in the characteristics of LSMPs associated with heat waves may be minor. Grotjahn (2015) shows trends in an LSMP index for California hot spells in reanalysis and both historical and future simulations by a climate model. Future changes in heat waves may also be modulated by changes in land–atmosphere interactions. For example, global warming may shift the climate regime northward and establish new transitional zones with strong land–atmosphere coupling strength (Seneviratne et al. 2006) and drying over the subtropics and mid-latitudes during summer in the future may increase land–atmosphere feedbacks in general (Dirmeyer et al. 2012) and result in amplified heat wave response to global warming. For CAOs, Schneider et al. (2015) and Gao et al. (2015) both demonstrated that the mean warming and the reduced temperature variance due to polar amplification account for most of the decrease in CAOs in a warmer climate. Gao et al. (2015) further found that changes in temperature skewness associated with changes in blocking frequency and thermodynamical modulation by melting of snow and sea ice along the 0 °C isotherm contribute additionally to regional differences in CAO changes.
Regional climate models have also been used to investigate temperature extreme changes in the future. Kunkel et al. (2010) showed that the regional model has superior skill in simulating heat waves compared to the global models, but it largely inherited the climate sensitivity from the GCMs. They found considerable regional variations in future heat wave characteristics depending on the emission scenarios and GCMs used. Using a suite of regional climate simulations driven by 5 ensemble members of a GCM, Diffenbaugh and Ashfaq (2010) found that hot extremes are intensified over much of the US associated with a summer anticyclonic circulation anomaly in the future, which reduces precipitation and soil moisture, and amplifies severe hot and dry conditions. The regional simulations showed a summer anticyclonic anomaly that is more widespread than in the GCM.
Overall, some robust changes in temperature extremes have emerged from analysis of the CMIP3 and CMIP5 multimodel ensembles as well as targeted global and regional climate modeling efforts. Model projections generally indicate increased frequency, intensity, and duration of heat waves, and despite global warming, cold extremes may persist in some regions due to changes in blocking events (Vavrus et al. 2006; Gao et al. 2015). More detailed analyses using multimodel ensembles may provide further insights on the mechanisms of heat wave changes associated with land–atmosphere coupling and potential changes in LSMPs affecting both heat waves and cold air outbreaks. Additionally, well-designed climate modeling experiments combined with observational analyses that target specific mechanisms and hypotheses may help isolate different factors and provide constraints for more robust projections of temperature extreme changes.
This review paper summarizes our current knowledge of, and context for developing new understandings about, extreme hot and cold temperature events affecting regions of North America. The topic of extremes encompasses many scientific issues and a breadth of time and space scales. Understanding extreme events ranges from how events are defined and measured, to how extreme events are studied statistically, theoretically, and with models. To reduce the breadth of scales, the paper focuses upon the large scale meteorological patterns (LSMPs) that accompany extreme temperature events (ETEs) of short duration. However, even with this narrower focus, one easily sees that much further progress is needed to understand the properties of extreme temperature events and their supporting LSMP processes. Thus, the primary goal of this report is to provide guidance to researchers interested in studying extreme temperature events, especially in relation to their associated LSMPs. As such, a variety of techniques are described for identifying the LSMPs, including the relative merits of each approach. A variety of analysis tools are identified which highlight the linkage of LSMPs to both remote low frequency phenomena and local factors. Also highlighted is the information gleaned from multiple climate models including simulation issues and projected trends.
Section 2 presents methods of defining extreme temperature events statistics using both simple indices and extreme value statistical techniques. Also surveyed are methods used to identify and connect LSMPs to extreme temperatures. Recent advances in statistical techniques, involving conditional extreme value analysis, offer an opportunity to connect LSMPs to extreme temperatures through appropriately defined covariates (i.e., indices of LSMPs) that supplement more straightforward analyses.
Section 3 surveys our current knowledge of LSMPs related to ETEs. Although phenomena ranging from synoptic-scale waves to planetary-scale climate modes are implicated as contributors to ETEs, existing information on (a) the physical nature of these contributions and (b) the dynamical mechanisms responsible for the implicated LSMPs is uneven and incomplete. A diagnostic formalism is put forth for systematically isolating the underlying physics of LSMP life cycles with an ancillary goal of identifying essential large-scale circulation “markers” for climate model validation purposes.
Section 4 summarizes the approaches used to model extreme temperatures, including dynamical methods using global and regional climate models and statistical models. Although climate models generally capture heat waves and cold air outbreaks with some fidelity compared with observations, they overestimate warm wave frequency and underestimate CAO frequency. Furthermore, while CMIP5 models properly represent many of the significant associations between ETEs and low frequency modes, they underestimate the collective influence of low-frequency modes on temperature extremes, particularly related to the Pacific Decadal Oscillation. Statistical methods used to relate LSMPs with extremes are limited mainly by small sample sizes of extreme events, so that only a small number of predictors can be included to ensure robust statistical relationships.
Section 5 surveys observed and projected trends in extreme temperatures. The studies that form the basis of the IPCC AR5 reports and the 3rd US National Climate Assessment do not consider the role of LSMPs in the magnitude and trends of extreme temperatures. However, a survey of the limited literature exploring these roles in North America and Europe suggests that future assessments, particularly at the regional scale, must include the connection between LSMPs and extreme temperature events.
In preparing this report it became apparent that much work remains to understand better past and future occurrences of ETEs and their underlying physical causes. Below are key questions relating to LSMPs that may guide future research.
Data and statistical methods Can statistical significance tests be developed for all LSMP detection methods? What are the effects of spatial resolution and gridding of station data on extreme value temperature analysis? How shall multivariate extreme statistical tools be optimally employed? What extreme values are most useful to various application sectors? Can the uncertainty be quantified in statistical quantities for given lengths of the observational or model data?
Synoptic behavior and physical mechanisms What physical/dynamical mechanisms form and maintain these rare LSMPs? Is more than one type of LSMP pattern responsible for a given regional ETE? Do specific LSMPs arise by different processes? (e.g. Lee and Grotjahn 2015) Can one quantify the extent to which low frequency phenomena such as blocking, MJO and ENSO contribute to each ETE? (e.g. Dole et al. 2014) Can one quantify the relative contributions of local factors such as topography, land surface states, and land use/land cover in ETE duration and severity? How do LSMPs vary with season and regional location? How sensitive is LSMP characterization to how the associated ETE is defined?
Model behavior and trends How well do global and regional climate models simulate the synoptic-dynamic behavior of relevant LSMPs and their variability and trends compared to observations/reanalyses? Can deficiencies in model dynamics or physical parameterizations be linked to model deficiencies in LSMP properties? How well must LSMP properties be simulated by global climate models to provide sufficiently accurate lateral boundary conditions for regional climate models to accurately simulate the associated ETE? Can metrics be designed to discern and understand differences in model skill in simulating an ETE and the associated LSMP?
Future properties What are the relative roles of dynamical and thermodynamical changes to future changes in LSMPs associated with ETEs? Considering current limitations of climate models, how can one optimally assess the uncertainty in model projections of future changes in relevant LSMPs and their ETEs? Beyond ensemble modeling, what type of coordinated experiments can be performed to systematically evaluate models and infer the sources of model differences in LSMPs and ETEs? How can such model evaluations improve our interpretation of model projections of future changes in extremes and LSMPs?
While many gaps exist, investigators now have a large variety of tools and a useful LSMP framework by which to pursue better understanding of both historical and future extreme temperature events.
The authors thank Dr. Christopher J. Paciorek for his assistance in preparing this article. The authors also thank Dr. Steven Vavrus for his comments. Most of the authors are members of the US CLIVAR Extremes working group who greatly appreciate the support provided by the US CLIVAR office. This report was enhanced by discussions held at the 2013 workshop on Analyses, Dynamics, and Modeling of Large Scale Meteorological Patterns Associated with Extreme Temperature and Precipitation Events held at the Lawrence Berkeley National Laboratory and also funded by US CLIVAR (https://usclivar.org/meetings/extremes-workshop-agenda). Research by Grotjahn was funded in part by NSF Grant 1236681 and also supported by the USDA National Institute of Food and Agriculture, Hatch project CA-D-LAW-4264-H. Leung and Wehner were supported by the Regional and Global Climate Modeling Program of the Office of Biological and Environmental Research in the Department of Energy Office of Science. Pacific Northwest National Laboratory is operated by Battelle Memorial Institute for the DOE under contract DE-AC05-76RL01830 and Lawrence Berkeley National Laboratory is under contract DE-AC02-05CH11231. Black was supported by the US Department of Energy, Office of Biological and Environmental Research, awards DE-SC0004942 and DE-SC0012554, and the National Science Foundation Grant ARC-1107384. Gutowski was supported by National Science Foundation Grant ARC1023369 and Department of Energy Grant DESC0006643. Gyakum was supported by a Natural Sciences and Engineering Research Council of Canada Discovery Grant, and by an International Polar Year Grant. The National Center for Atmospheric Research is sponsored by the National Science Foundation.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.