Abstract
Baseline assumptions play a crucial role in conducting consistent quantitative policy assessments for dynamic Computable General Equilibrium (CGE) models. Two essential factors that influence the determination of the baselines are the data sources of projections and the applied calibration methods. We propose a general, Bayesian approach that can be employed to build a baseline for any recursivedynamic CGE model. We use metamodeling techniques to transform the calibration problem into a tractable optimization problem while simultaneously reducing the computational costs. This transformation allows us to derive the exogenous model parameters that are needed to match the projections. We demonstrate how to apply the approach using a simple CGE and supply the full code. Additionally, we apply our method to a multiregion, multisector model and show that calibrated parameters matter as policy implications derived from simulations differ significantly between them.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Nowadays, Computable General Equilibrium (CGE) models are considered the workhorse models of policy analysis focusing on economywide effects induced by exogenous economic shocks or policy interventions (de Melo 1988; Shoven and Whalley 1984; Dixon and Jorgenson 2013). For example, CGE models have been widely used to assess the impacts of policies in the area of international trade (Hertel 1997; Hertel et al. 2007; Caliendo and Parro 2014), migration (Stifel and Thorbecke 2003; Fan et al. 2018), agricultural policies (MilczarekAndrzejewska et al. 2018; Taylor et al. 1999), and energy(Burniaux and Chateau 2014) or climate policies (Böhringer et al. 2014; Fujimori et al. 2016; Böhringer et al. 2021).
However, longstanding criticisms of CGE models include that these models have weak econometric foundations (McKitrick 1998; Jorgenson 1984). This criticism arises from the fact that CGE models are complex, and the available empirical data is limited, implying that it is often impossible to estimate all model parameters econometrically (McKitrick 1998; Jorgenson 1984; Hansen and Heckman 1996). Thus, typically model parameters are either assumed ad hoc or weakly derived from empirical data to calibrate CGE models.
In static CGE models, a classical baseline calibration corresponds to calculating model exogenous variables, such that model output in the equilibrium replicates the economic structure defined by a given social accounting matrix (SAM) empirically observed in a specific base year. This approach can be problematic because generally, an infinite number of parameter setups can be generated such that all of them can perfectly replicate the SAM. A good case in point is the oftenused practice of assuming ad hoc values for relevant elasticities of substitution or transformation to determine remaining parameters of corresponding Constant Elasticity of Substitution (CES) and Constant Elasticity of Transformation (CET)functions. Thus, calibration of static CGE models corresponds to a more or less arbitrary parameter specification. In response to this criticism, Systematic Sensitivity Analysis (SSA) is being increasingly used in CGE model applications. With this approach endogenous CGE output variables are simulated based on sampled CGE model parameters that are derived from estimated or assumed distributions (Olekseyuk and SchürenbergFrosch (2016)).
However, though SSA is a good method to reveal induced uncertainty of model outputs explicitly as it derives the means and variances of all relevant endogenous variables based on an assumed distribution of parameters, it does not provide a way to use external information to “update” the assumed distribution. Hence, it is not an appropriate procedure to reduce model uncertainty. Furthermore, SSA does not provide a method to integrate observed model uncertainty into the derivation of optimal policy choices.
For dynamic CGE models, in addition to the base year calibration, modellers also need to calibrate the longterm behaviour (Sánchez 2004). Typically, while constructing the baseline of a dynamic model, modellers use historical data and forecasts to specify relevant model parameters minimizing forecasts errors comparing simulated model output with observed or forecasted data over the entire forecast period from the base year to the end year. Calibration of a dynamic baseline has the advantage of focusing on the longterm behaviour of the model, i.e., delivering information about the key parameters that drive the model response to an exogenous shock. Understanding the response of the economy to exogenous shocks is not only crucial to predict correctly future economic development, such as sectoral growth, employment, capital stock, trade or Greenhouse Gas (GHG) emissions in the baseline but also to assess counterfactual economic developments implied by different policy scenarios (Nong and Simshauser (2020); Pothen and Hübler (2021), see also Sect. 5.3). Moreover, the different approaches applied to derive model baselines and calibrating model parameters are far from being innocuous but rather crucially impact on corresponding policy analyses. Accordingly, parameter calibration and baseline derivation has received wideattention for several types of numerical models that are used for policy analysis (e.g., Dynamic Stochastic General Equilibrium (DSGE) models in Gomme and Lkhagvasuren (2013), New Quantitative Trade (NQT) models in Pothen and Hübler (2021) and spatial trade models in Paris et al. (2011)).
Technically, baseline calibration of dynamic equilibrium models corresponds to a high dimensional optimization problem, i.e., a set of model parameters must be identified such that the equilibrium values of the variables of interest should match the forecasts from external sources of data. Additionally, a set of theoretical restrictions like closure rules also need to hold valid on the model parameters. Given the high dimensionality of parameters and output variables, solving this problem is very challenging. Moreover, beyond the theoretical parameter restrictions, a priori expert information regarding the empirical range of model parameters is generally available, and prior parameter distributions can formally represent this. Hence, given all these constraints, a Bayesian estimation approach appears to be an appropriate methodological framework for baseline calibration of dynamic equilibrium CGE models. Moreover, Bayesian estimation techniques are applied to estimate parameters of DSGE models (Smets and Wouters 2003; Hashimzade and Thornton 2013). In particular, like CGEs also DSGE models correspond to sound microfounded and theoretically consistent model structures, where applying Bayesian estimation techniques DSGE models combine these theoretical properties with good forecasting performance. For example, Smets and Wouters (2003) have demonstrated that the estimated DSGE models perform quite well in forecasting compared to standard and Bayesian vector autoregressions. Not only does the systemwide estimation procedure deliver a more efficient estimate of the structural model parameters, but it also provides a consistent estimate of the structural shock processes driving recent economic developments. Understanding the contribution of the various structural shocks to recent economic developments is an important input in the monetary policy decision process. Accordingly, they conclude especially that applying Bayesian estimation techniques to calibrate DSGE models the latter can become a useful tool for projection and policy analysis in central banking. However, compared to CGE models DSGEs are still highly aggregated models, which assume highly aggregated economic agents, e.g., do not include heterogenous sectoral structures. Therefore, in contrast to DSGE models applying Bayesian estimation techniques to estimate parameters of a CGE model is still challenging due to the much more complex structure and much high number of parameters of the latter.
Some researchers have used an informal Bayesian estimation procedure by which they run a model forward over a historical period and compare model results with historical data (Gehlhar 1994; Kehoe et al. 1995; Dixon et al. 1997). Though such approaches can be helpful to revise parameter estimates and recalibrate the model (Tarp et al. 2002), they are nevertheless ad hoc and do not yet offer a systematic Bayesian procedure that can be applied baseline calibration of a dynamic model. Alternatively, Arndt et al. (2002) and more recently Go et al. (2016) propose a maximum entropy approach for parameter estimation of CGE models by using information theory to estimate CGE parameters based on a sequence of observed SAMs. Their approach can be interpreted as a special case of the Bayesian approach and hence, already provides several advantages as compared to the standard calibration methods. However, the approach still has limitations that can be addressed in a general Bayesian framework. First, it is based on specific assumptions regarding a priori parameter distributions, which can be significantly relaxed in a general Bayesian framework (Heckelei and Mittelhammer 2008). Second, this approach is focused on an expost analysis, and thus, it requires empirical observations of corresponding SAMs. Furthermore, so far, the Cross Entropy (CE)method has only been applied to single country and static CGE models, and this approach cannot be easily extended to baseline calibration of dynamic CGE models.
Our paper proposes a Bayesian estimation approach that generalizes the CEmethod of Arndt et al. (2002); Go et al. (2016) thereby allowing a more direct and straightforward formulation of available prior information (see Sect. 2). An important innovative step in our approach is that we complement the Bayesian estimation with metamodeling techniques (Kleijnen and Sargent 2000) to replace the CGE model with a simplified surrogate model to reduce complexity and thereby, significantly reduce the computational effort (Sect. 2.1). Our Bayesian approach also enables us to simulate endogenous CGE outputs based on sampled model parameters that are derived from the corresponding posterior distribution, where technically metamodeling also facilitates MetropolisHasting sampling from this posterior distribution. We show how to apply the approach with a toy model and provide the complete code for this demonstration (Sect. 3). Since our proposed approach can also be applied to calibrate complex multiregion and multisector CGEs subsequently in (Sect. 4) we demonstrate our approach by calibrating the dynamic baseline of the Dynamic Applied Regional Trade (DARTCLIM) model. Lastly, we also show that the chosen calibration approach has a noticeable impact on policy assessments by using our method to construct two baselines for DARTCLIM albeit with differences in the extent of calibration (Sect. 5).
2 Methods
Simplified, the application of CGE models in policy analysis follows three main steps in general (Fig. 1).
The first step is collecting statistical data, like SAMs or Input/Output (I/O)tables besides information about exogenous parameters (like elasticities) and forecasts for key outputs from other studies. The second step is model calibration, where the parameters of the CGE are derived to replicate observed data and forecasts to a certain degree. When the calibration gives a satisfying specification, the model is applied in policy analysis by implementing policy shocks, and the policy impact is measured by comparing outputs from the policy scenario to the baseline scenario. In the rest of this section, we present the foundations of calibration and derive a consistent and comprehensive approach to the calibration step.
Formally, let F denote a CGE model, which implicitly determines outputs, \({\varvec{y}}\) as a function of a set of model parameters, \({\varvec{\theta }}\):
F is an Idimensional vector valued function, \({\varvec{y}}\) an Idimensional vector of endogenous output variables and \({\varvec{\theta }}\) a Kdimensional vector of exogenous model parameters. Model parameters can be further disaggregated into different subsets, e.g. behavioural parameters or exogenous variables. The latter also includes policy variables that are controlled by the government, e.g. taxes or tariffs, etc. Further, exogenous variables include demographic or economic variables uncontrolled by the government, e.g. world market prices, population growth, etc.... Changes in the values of exogenous variables correspond to exogenous shocks, which induce changes in the endogenous variables. The latter corresponds to the response of the economic system to exogenous shocks. Behavioural parameters define the response of the economic system to exogenous shocks. Hence, assuming behavioural parameters correspond to their true values implies that the model replicates the true response of the economy to exogenous shocks. In this context, we define a dynamic model as a model for which variables and parameters are defined for different periods, while a static model corresponds to a model for which variables and parameters are only defined for one specific time (base) period. Further, we define baseline by a complete parameter specification which corresponds to a clearly defined base run scenario. Obviously, since the model \(F({\varvec{y}},{\varvec{\theta }})\) determines for any complete parameter specification (\({\varvec{\theta }}\)) the corresponding endogenous output variables a baseline also delivers specific values for all endogenous variables (\({\varvec{y}}\)).
An advantage of CGE models is that one can simulate counterfactual scenarios, i.e. one can calculate the values of the endogenous variables that result from assuming parameter values that differ from their corresponding baseline values. Hence, one can simulate the impact of different policies by changing the subset of policy variables compared to their baseline values, or one can simulate the impact of different framework conditions on policy impacts by changing the subset of exogenous variables uncontrolled by the government. Finally, one can also simulate the impact of characteristic properties of the economic system on the impact of specific policies via simulating policy impacts, assuming different values for the subset of behavioural variables and comparing these to policy impacts derived for the baseline values. However, important for any modelbased policy analysis is that the baseline parameters can be empirically validated. In particular, this implies that the subset of behavioural parameters corresponds to the true structural properties of the economic system.
Given the fact that in contrast to endogenous and exogenous variables, behavioural parameters can generally not be directly observed, empirical model validation is based on a comparison of observed values of the endogenous model variables with corresponding values derived from model simulations. Accordingly, all model validation methods basically correspond to the identification of behavioural model parameters that maximize the likelihood of the model specification. Given enough observed data, \({{\varvec{y}^o}}\), of sufficient quality, \({\varvec{\theta }}\) can be estimated econometrically. In the context of CGEs, data is usually scarce, and therefore, the identification of \({\varvec{\theta }}\) becomes a calibration problem. A standard calibration method for CGEs is based on given empirical observations of endogenous and exogenous variables in a given base year, where behavioural parameters are determined in a way that the observed variables are replicated in this base year. Since for this procedure, data is only observed in a single (base) year, we call this a static baseline calibration method. In contrast, assuming parameters are determined based on variable values observed for a sequence of time periods, we use the term dynamic baseline calibration method.
However, as we already explained above, a problem of a static baseline calibration corresponds to the fact that there exists an infinite set of different behavioural parameter specifications that imply that simulated outputs perfectly replicate a given set of observed variables, while simulated responses of the economic system to given exogenous shocks differ in contrast significantly for different behavioural parameter values. Therefore, modelbased policy analysis becomes arbitrary. We suggest a Bayesian framework for an advanced parameter calibration method to avoid this arbitrariness. The latter can be applied as a static as well as a dynamic baseline calibration method.
In the general Bayesian framework, observed variables are noisy, i.e., data \({\varvec{y}^o}= \{{Y^o}_1, \dots , {Y^o}_N\}\) correspond to true variable values, \({\varvec{y}}= \{{{Y}}_1, \dots , {{Y}}_n\}\) and noises \({\varvec{\epsilon }}= \{{\epsilon }_1, \dots , {\epsilon }_N,\}\). Assuming \({\epsilon }_n\) is iid normally distributed, \({\epsilon }_n \propto N(0,I)\) implies that the posterior results as:
Building upon this general Bayesian framework, we develop a calibration procedure for quasidynamic CGE models. To this end we define different subsets of parameters, i.e., \({\varvec{\theta }}= ({\varvec{\theta }}_0,{\varvec{\theta }}_t)\). \({\varvec{\theta }}_0\) denotes parameter values in the base year while \({\varvec{\theta }}_t\) denotes parameter values in a period \(t=1, \dots ,T\) different to the base year. Given a forecast for a subset of output variables, \({\varvec{z}^o}\in {\varvec{y}}\), and \({\varvec{y}^o}_0 \in {\varvec{y}}\) empirical observations of a subset of CGEvariables in the base year \(t_0\), we define \({\varvec{\epsilon }}_0 = {\varvec{y}^o}_0  {\varvec{y}}_0\) and \({\varvec{\epsilon }}_z = {\varvec{z}^o} {\varvec{z}}\), and \({\varvec{\epsilon }}= ({\varvec{\epsilon }}_0, {\varvec{\epsilon }}_z)\). Additionally, assuming normal distributions for \({\varvec{\epsilon }}\propto N(0, \Sigma _{{\varvec{\epsilon }}})\) and \({\varvec{\theta }}\propto N(\bar{\varvec{\theta }}, \Sigma _{{\varvec{\theta }}})\) with the covariance matrices \(\Sigma _{{\varvec{\epsilon }}}\), \(\Sigma _{{\varvec{\theta }}} = (\Sigma _0, \Sigma _T)\) as diagonal matrices with elements \(\sigma ^2_{{\epsilon }}\), \(\sigma ^2_{{\theta }}\), we can derive the following optimization problem for the Highest Posterior Density (HPD) estimator:
\(H({\varvec{\theta }}_0, {\varvec{\theta }}_T)\) takes additional parameter constraints into account that are induced by economic theory or apriori expert information. In general, given observations \({\varvec{y}^o}_0\) and forecasts \({\varvec{z}^o}\), HPD estimation follows an optimization problem described by the system 3. It is also possible to choose other distributions, i.e., other extremum metrics, while keeping the general idea the same. Moreover, please note that in general, the approach 3 includes the option that some variables or parameters are known with certainty, e.g. governmental policies like taxes or tariffs are normally known with certainty. In this case, these variables or parameters are fixed to their prior values and excluded from the prior density function.
However, outputs \({\varvec{z}}\) are only defined as an implicit function by theCGE, and therefore, the solution of the optimization problem is very tedious. Hence, complex methods of simulation optimization (SO) have to be applied to solve (3). We apply metamodeling techniques to reduce complexity and computational effort (see 2.1 for a short introduction). The advantage of using metamodels is that they may result in more efficient SO methods. As a result, F is substituted with a metamodel M, which approximates the mapping between model parameters and a subset of output variables, \(({\varvec{y}}_0, {\varvec{z}})\), derived from the original model F.
The formulation in (3 and 4) allows a straightforward interpretation of the assumed variances \(\sigma ^2_{{\epsilon }}\) and \(\sigma ^2_{{\theta }}\). In particular, we can interpret \(\sigma ^2_{{\epsilon }}\) as weights that indicate the importance of matching the corresponding forecast \({Z^o}\) relative to the other variables. A lower value of \(\sigma ^2_{{\epsilon }}\) suggests higher importance relative to \(\sigma ^2_{{\epsilon }'}\). The variance \(\sigma ^2_{{\theta }}\) captures the information we know about a parameter \({\theta }\) with a lower value for \(\sigma ^2_{{\theta }}\) compared to \(\sigma ^2_{{\theta }'}\) implying more certainty or higher knowledge about \({\theta }\) than \({\theta }'\).
2.1 Metamodeling
Metamodeling techniques are widely used in a variety of research fields such as design evaluation and optimization in many engineering applications (Simpson et al. 1997; Barthelemy and Haftka 1993; SobieszczanskiSobieski and Haftka 1997), as well as in natural science (Razavi et al. 2012; Gong et al. 2015; Mareš et al. 2016). In recent years, metamodeling is increasingly being applied to economic research. For example, Ruben and van Ruijven (2001) have applied the approach to bioeconomic farm household models to analyze the potential impact of agricultural policies on changes in land use, sustainable resource management, and farmers’ welfare; VillaVialaneix et al. (2012) have compared eight metamodels for the simulation of \(N_2O\) fluxes and N leaching from corn crops; Yildizoglu et al. (2012) have applied the technique to two wellknown economic models, Nelson and Winter’s industrial dynamics model and Cournot oligopoly with learning firms, to conduct sensitivity analysis and optimization respectively. Regardless of the research fields, the metamodeling technique simplifies the underlying simulation model, leading to a more indepth understanding. The technique also brings the possibility of embedding simulation models into other analysis environments to solve more complex problems, such as the previously described calibration process.
The use of metamodeling entails three steps: selection of metamodel types, Design of Experiments (DOE), and model validation (Kleijnen and Sargent 2000).
2.1.1 Metamodel Types
Metamodels are classified into parametric and nonparametric models (Rango et al. 2013). Parametric models, such as polynomial models (Forrester et al. 2008; Myers et al. 2016), have explicit structure and specification. On the other hand, nonparametric models do not depend on assumptions of model specification and determine the (I/O) relationship of the underlying simulation model using experimental data. Examples of nonparametric models consist of Kriging models (Cressie 1993; Yildizoglu et al. 2012; Kleijnen 2015), support vector regression models (Vapnik 2013), random forest regression models (Breiman 2001), artificial neural networks (Smith 1993), and multivariate adaptive regression splines (Friedman et al. 1991).
In this paper, we focus on the polynomial models that are defined by their order. For example, a secondorder polynomial model is given as follows:
where \({\theta }_1,....,{\theta }_k\) are the k independent variables, \({{Y}}\) is the dependent variable and \(\eta \) is the error term. The corresponding coefficients \(\beta \) are usually estimated through a linear regression based on least squares estimation (Chen et al. 2006). Polynomial metamodels generally provide only local approximations. Theoretically, certain types of metamodels like Kriging models (Kleijnen 2015) can also provide a global fit. Additionally, different metamodels may also be combined into an ensemble (BartzBeielstein and Zaefferer 2017; Friese et al. 2016). Some of the advantages of the polynomial models are:

They have simple forms that are easy to understand and manipulate.

They require low computational efforts.

They can be easily integrated into other research frameworks.
For a more thorough introduction to other types of metamodels, see for example Dey et al. (2017); Simpson et al. (2001).
2.1.2 Design of Experiments
To utilize the metamodels, we need to estimate the corresponding coefficients. We generate the simulation sample by DOE, which is a statistical method of drawing samples in computer experiments (Dey et al. 2017) and perform the estimation by entering the simulation sample into the simulation model. DOE could be setup in two ways: the classical experimental design and the spacefilling experimental design (see Fig. 2). The former places the sample points at the boundaries and the centre of the parameter space to minimize the influence of the random errors from the stochastic simulation models. However, Sacks et al. (1989) have argued that this is not the case for deterministic simulation models where systematic errors prevail. Therefore, the spacefilling experimental designs should be employed to replace the classical ones. Among popular spacefilling designs, Latin Hypercube design enjoys great popularity due to its ability to generate uniformly distributed sample points with ideal coverage of the parameter space as well as the flexibility with the number of the sample points (Sacks et al. 1989).
2.1.3 Model Validation
Validation refers to assessing whether the prediction performances of the metamodels hold an acceptable level of quality (Kleijnen 2015; VillaVialaneix et al. 2012; Dey et al. 2017). Normally, two samples are needed to assess the quality of a derived metamodel: the training sample and the test sample. The training sample is used to fit the parameters of the metamodel, whereas the test sample is used to validate the trained metamodel, and the test sample must include data points that are not part of the training sample. It is important that the metamodels have good predictions while maintaining generality. For this reason, a test sample is essential because it helps us evaluate if the metamodels can be generalized and whether the simulation model can be replaced with them.
The following textbook statistics are often considered to assess the validation results.
where \({{Y}}_i\) and \({Y^o}_i\) are the predicted values and true values for the test sample at sample point i, and \({\overline{{Y^o}}}\) is the mean of \({Y^o}\) in the test sample. In regression analysis, \(R^2\) is a statistical measure of how close the data are to the fitted regression line. The root mean squared error (RMSE) is the square root of the variance of the residuals. It indicates the absolute fit of the model to the data i.e., how close the model’s predicted values are to the true values.
To compare the prediction performances for dependent variables that have different scales, we introduce the absolute error ratio (AER), which is calculated by taking the absolute value of RMSE divided by the corresponding mean:
The AER metric gives us an idea of how large the prediction errors are in comparison to the true simulated values on average, i.e., the lower the AER values, the better the prediction performances.
3 Demonstration
In this section, we first demonstrate how the theoretically derived approach (see Sect. 2) can be applied using a simple toy model (Hosoe et al. 2010, 2016). We use the toy model to provide technical insights to the reader on how our approach can be applied. Therefore, we also refer explicitly to the corresponding programming code of our calibration procedure, giving pointers on where each step is implemented. The complete implementation code can be found as a supplement to this paper^{Footnote 1}. The DARTCLIM application in Sect. 4 is used to demonstrate that our approach can also handle the calibration of huge complex models with a very large number of parameters (\(\gg 1000\)).
The toy model we use is from the General Algebraic Modeling System (GAMS) model library (GAMS Development Corporation 2020), and is a simple four sectors, two factors, one household recursive dynamic CGE model (Hosoe et al. 2010, 2016).
In order to apply our dynamic calibration procedure, five main steps are necessary, which are summarized in Algorithm 1. We implemented the individual steps in a mix of R (R Core Team 2020) and GAMS (GAMS Development Corporation 2020), where the GAMS part only requires a free demo license^{Footnote 2}. The R code drives the full example, where two scripts have to be run to replicate the steps and results described in the following (first R/sample_generation.R and second R/metamodels.R). The computation should be done in about five minutes on a current system.
Following the standard calibration procedure of CGE models, an observed SAM is used to calibrate behavioural model parameters. However, based on an observed SAM using equilibrium conditions of the CGE the behavioural parameters are underdetermined, i.e. for any observed SAM there exists infinite parameter values for which equilibrium conditions are fulfilled. Accordingly, a subset of parameters is usually exogenously fixed, and the remaining parameters are calculated using equilibrium conditions based on an observed complete SAM.
Similar to Hosoe et al. (2010, 2016) we exogenously determine the parameters listed in Table 1, i.e. the four elasticities of substitution for each of the Armington function, \(\textit{sigma}\), the four elasticity of transformation for each of the CETfunction, \(\textit{psi}\), as well as the exogenous variables population growth, \(\textit{pop}\), rate of return for capital, \(\textit{ror}\), depreciation rate, \(\textit{dep}\). Further, exogenously determined parameters include a parameter determining sectoral investments, i.e. the elasticity of investment, \(\textit{zeta}\)^{Footnote 3}.
While the standard calibration procedure implicitly assumes that the true SAM can be completely and perfectly observed, empirically observed SAM data is often noisy or incomplete. Hence, like the behavioural parameters, also the observed SAM data correspond to stochastic variables. In general, a Bayesian approach can deal with partly noisy and incomplete SAM data (see also Go et al. (2016)). For example, forward and backward linkages of a sector or detailed data on specific factor inputs might be missing, or (I/O) relations might be noisy, i.e. they do not perfectly match with the true SAM. Dealing with this kind of incomplete SAM data CGEmodellers normally apply a data imputation procedure like, for example, the entropy approach suggested by Robinson et al. (2001). Following the latter missing data is estimated based on a prior SAM. The latter is constructed using available statistical data taken from SAMs available for prior periods, or data available from other countries with a similar economic structure like the country to be analyzed or data taken from expert estimations. However, imputed SAMs crucially depend on the constructed prior SAM where generally different options exist to construct this prior. Hence, constructed SAM data is noisy. As a way to deal with this, we suggest the following: We assume that based on available data one can construct a finite set of different idealtypical SAMs, \(\{{\varvec{SAM_r}} \}\), where the true SAM follows as a convex combination of these idealtypical SAMs. Let \(\omega ^{\text {true}}\) denote the vector of weights of the idealtypical SAMs, i.e. it holds: \(SAM^{\text {true}} = \sum \nolimits _r \omega _r SAM_r\). Thus, once the set of idealtypical SAMs have been constructed, we simply include the weights \(\omega \) into \({\varvec{\theta }}\).
For the following demonstration, we constructed two SAMs, where each corresponds to a specific economic structure characterized by specific forward and backward linkages of the agricultural sector as well as by a specific labour intensity of agricultural production (see gams/sam.gms#L*). The true SAM corresponds to a linear combination of these two SAMs, where the SAM combination parameter \(\omega \) is a priori unknown and hence randomly drawn between zero and one.
Further, to be able to evaluate the goodness of fit of our calibration procedure we generate a benchmark parameter specification \({\varvec{\theta }}^{\text {true}}\) (see R/sample_generation.R#L3233. Based on this benchmark specification, we run the model for \(T=30\) time periods generating a time path for all endogenous variables. However, given the fact that in a true empirical application of our approach, exogenous forecasts will be available only for a limited subset of endogenous variables let \({\varvec{z}}_t = \{{{Z}}_{k,t}\}\) denote a matrix of selected endogenous variables of the CGE, where \({{Z}}_{k,t}\) denotes the forecasted equilibrium path of the endogenous variable k observed for time periods t.
3.1 Design of Experiments
The first step in the process is the generation of a suitable sample to estimate and validate the metamodels.
After having decided on the parameters to be used for the calibration, the allowed range for the parameters is defined based on prior experience, expert opinions, and literature review. For ror, dep, pop, zeta we picked a small range around the original values (\(\pm 0.01\) or \(\pm 0.02\)). For the two elasticity parameters sigma and psi we picked a large range of [1, 18] (see R/sample_generation.R#L9 17). This information is then transformed into a suitable format for the Latin Hypercube Sample (LHS) generation function, and a larger estimation sample and a validation sample is generated.
The number of necessary simulations depends on the kind and form of the metamodel to be estimated. A rule of thumb for the required simulations is the number of coefficients times ten for polynomial models. As we will apply a secondorder polynomial model, this is \(\approx {\varvec{\theta }}^2 \times 10 = 13^2 \times 10 = 1690\). For the validation sample, which is separate to the estimation sample, we generate \({\varvec{\theta }}^2 \times 2 = 338\) parameter settings (R/sample_generation.R#L2831).
The R code exports the sample to a GDXfile(see #L46) and calls the code for the next step (see #L4160).
3.2 Simulations
Following the sample generation, one needs to calculate the individual simulations and collect the results. We abstain from calculating it in a cluster environment given the simple structure and, therefore, nearly instantaneous solves of the toy model. In order to calculate the results for the samples, minimal changes are necessary to the model: The parameter values need to be loaded from a GDXfile (as is often the case for nontoy models) and unload the simulation results in the end. In our case, we also replaced the SAM, that was entered in textform into the model (see gams/cge/model.gms#L4445,208). A second part needed is some wrapping code that loops over the individual parameter settings and collects the results, which we also implemented in GAMS (see gams/mm_simulations.gms#L*). This code defines the necessary parameters(see #L469), loops over all parameter settings(see #L7190), scales results for numerical stability(see #L9297) and finally unloads the aggregated results for further consumption.
The loop in lines #\(7190\) also provides the basis for splitting the work across multiple machines. Instead of looping over the full set of specifications, only loop over a subset of those, save the partial results and aggregate those together in a final step. This is possible since the individual specifications do not depend on each other.
3.3 Metamodels
Using the results from the previous two steps, we can now estimate the metamodels, that replace the original CGE \(F({\varvec{y}}, {\varvec{\theta }}) \equiv 0\) with a set of metamodels \({\varvec{z}}= M({\varvec{\theta }})\) (see R/metamodels.R#L*). In our simple toy application we assume \({\varvec{y}}= ({\varvec{y}}_0, {\varvec{z}})\), where \({\varvec{z}}\) corresponds to the outputs being matched in the calibration procedure (see Table 1), and \({\varvec{y}}_0 = {\varvec{y}}\setminus {\varvec{z}}\) to other endogenous model outputs including SAM entries in the base year. To keep our demonstration example simple, we estimate metamodels for each target \({{Z}}_j \in {\varvec{z}}\) only for the base year \(t_0\) and the end year \(t_{30}\)^{Footnote 4}.
Given the relatively simple structure of the underlying CGE and the relatively small number of parameters, we can fit a secondorder polynomial model that captures the relationships quite well. In a more realistic setting with a larger model that requires much more computational effort, it will become expensive to run the required number of simulations (see Sect. 4 for a way to deal with this). The required number of simulations roughly scales quadratically with the number of calibration parameters. We use linear regression to estimate the coefficients of the metamodels. With the help of some loops and data wrangling of the estimation and validation samples, we can estimate the metamodels and calculate the validation measures (see R/steps/estimate_models.R#L*). The estimations and validations use helper functions from the R packages DiceDesign / DiceEval (Dupuy et al. 2015).
As evidenced by the low AER, we achieve a nearly perfect fit for all of the selected endogenous output variables, \({\varvec{z}}\), in the final year \(t_{30}\). For comparison we also show the mean, the RMSE and \(R^2\) values in Table 2. For brevity, we skip the results for the base year \(t_0\), where the achieved goodness of fit for base year variables is similar to the presented results.
3.4 Bayesian Calibration
With the help of the estimated metamodels, we can now solve the calibration problem (see Equation 4). We implemented the calibration procedure in GAMS, but R (or any other language with the possibility to solve NLPs problems) would also be fine (see gams/calibration.gms#L*). The model for the optimization problem is implemented in GAMS (see gams/calibration/model.gms#L*), which consists of the objective function ((see #L90118) and the metamodels (see #L120160)^{Footnote 5}. The required data, metamodel coefficients, parameter bounds, variances and priors, and forecasts are exported in the metamodel estimation code (see R/metamodels.R#L1516.) As shown in Table 5 we assumed large values for the variances, with their values being set directly in the code (see gams/calibration.gms#L2331). Please note that we include only a subset of time periods (only the base year and the end year) into the Bayesian estimation to account for the fact that often forecasts for relevant variables are only available for selected time periods. The GAMS code for the calibration is called automatically from R (see R/metamodels.R#L1840).
Please note further that in our simple toy model application, we generally assume that behavioural parameters are constant over time, i.e. \({\varvec{\theta }}_t={\varvec{\theta }}_0\)^{Footnote 6}. Second, please note that while \({\varvec{\epsilon }}_z = {\varvec{z}^o} M({\varvec{\theta }})\), it holds \({\varvec{\epsilon }}_0 = SAM^o  \sum \limits _r \omega ^{*}_r SAM_{r}\). Hence, in general observed values, \(SAM^o_{i,j}\) may differ from the estimated true SAMvalues, \(\sum \limits _r \omega ^{*}_r SAM_{r,i,j}\).
Overall, the Bayesian approach allows to combine multiple sources of information: (1) empirical observation of SAMvalues in the base run, \(SAM^0_{ij}\), (2) dynamic forecasts of endogenous variables, \({\varvec{z}}\) as well as expert information on economic structures, captured in derived idealtypical SAMs \(SAM_r\) and their corresponding prior weights, \(\omega _r\). Finally, please note that beyond dealing with noisy SAM data, the Bayesian approach can easily deal with missing SAMdata, where \(\sum \limits _r \omega ^{*}_r SAM_{r,i,j}\) correspond to the Bayesian estimation of missing SAMvalues.
3.5 Evaluation
Using the found parameter values \({\varvec{\theta }}^*\) and given that we know the “truth” in this example, we can evaluate the model predictions in three aspects: (1) The first one is using a (gof) measure, which is proportional to the (negative) likelihood part (\({{\mathcal {L}}({\varvec{y}^o}\mid {\varvec{\theta }})}\)) of our target function (see Equation 3). This captures how modelers view the importance of the different targets in the calibration procedure, which is reflected by the relative value of \(\sigma ^2_{{{Z}}_j}\) to \(\sigma ^2_{{{Z}}_{j'}}\), but also \(\sigma ^2_{{{Z}}_j}\) to \({{Z}}_j\). A lower value for goodnessoffit gof means a higher likelihood and hence a better fit, with a value of zero indicating a perfect fit. (2) We can compare model outputs of the metamodel with simulation outputs of the true model. (3) We can compare estimated parameters \({\varvec{\theta }}^*\) with their corresponding true parameter values \({\varvec{\theta }}^{truth}\). In the Tables 3, 4 and 5 we report different gof measures. These Tables are created as the final step in the R code (see R/metamodels.R#L43).
As evidenced by the results in Table 3 comparing predicted outcomes of the metamodel and the calibrated CGE with the true outcomes for the forecasted variables we achieve an excellent fit^{Footnote 7} for the calculated gof measure presented in Table 3. Further, we note a difference between the gof from the calibration step (calculated comparing the true and the predicted outcomes of the metamodel), and the gof calculated for the prediction of the calibrated CGE (\({\varvec{\theta }}^*\)).
So far, we only looked at the targets that were active in the previous step. Next, we look at the percentage differences for all targets reported in Table 4 and we notice only small differences between the values predicted by the metamodels (M) and the values calculated using the true simulation model (F). The differences are below one per cent, with few exceptions and a maximum difference of 3.6 per cent. For the comparison between the true specification (\({\varvec{y}}^{\text {true}}\), with \(F({\varvec{y}}^{\text {true}}, {\varvec{\theta }}^{\text {true}}) \equiv 0\) and the values from the simulation model (\({\varvec{y}}^{F,*}\) with \(F( {\varvec{y}}^{F,*},{\varvec{\theta }}^*) \equiv 0\) we see few larger differences (\({\varvec{Z3_{LMN,30}}}\), \(Z3_{AGR,30}\)), but overall we see a very good match of values, even for those targets, that were not active in the calibration step.
Beyond prediction errors, it is interesting to analyze the goodness of fit regarding the estimation of the CGE parameters. As can be seen from Table 5 for most parameters, the true value could be nearly identified, with over \(60\%\) of the parameters having only a deviation below \(15\%\). The largest deviations can be found for the elasticity of substitution and elasticity of transformation for the sector LMN, with estimation errors of \(40\%\) and \(258\%\). Especially the latter corresponds to the high forecast error found for exports of the LMN sector. Interestingly, based on our approach, we can identify the true underlying economic structure, i.e. the Bayesian estimation delivers a posterior value for \(\omega \) of 0.41, which moves the uninformative prior value of 0.5 significantly towards the true value of 0.38, implying only an estimation error of slightly over \(8\%\).
Overall, based on achieved gof we can conclude that our dynamic calibration procedure is an efficient approach to estimate relevant CGEparameters, which corresponds to a superior method when compared to standard CGEcalibration. The latter follows especially given the fact that our suggested procedure allows an explicit evaluation of the goodness of fit of derived parameter estimations.
4 DART Application
In a second step, we will apply our approach to calibrate the DARTCLIM model to demonstrate that our approach is suitable to calibrate large multiregionsmultisector recursivedynamic CGE models.
4.1 Dynamic Applied Regional Trade Model
DARTCLIM is a recursive multiregion, multisector CGE model that is developed at the Kiel Institute for the World Economy and is being used to assess the effects of climate policies on the global economy (Winkler et al. 2021; Thube et al. 2022) The model is based on the Global Trade Analysis Project (GTAP)9 database (Aguiar et al. 2016), with an aggregation to 20 regions and 24 sectors. Detailed definitions on regional and sector aggregation are shown in Tables 9 and 10, respectively.
For the application of our calibration method, as shown in Table 6, we selected six regionspecific endogenous variables \({\varvec{z}}\) (total of 120 output variables) that need to be matched to their forecasts. The data for the projection of Gross Domestic Product (GDP) is taken from the OECD macroeconomic forecasts (OECD 2019) and the data for the electricity production and \(CO_2\) emissions comes from the World Energy Outlook (International Energy Agency 2018). Specifically, we will match the Compound Annual Growth Rate (CAGR) of the model outputs to those calculated from the forecasts for better numerical stability, assuming an exponential growth path.
Table 6 also shows the nine input parameters of DARTCLIM that we will use to replicate the forecasts of the above said endogenous variables in the baseline scenario. All model parameters are either regionspecific and/or sectorspecific and thus, account for a total of about 1500 input variables. These nine input parameters correspond to the typical set of model parameters chosen by modellers to calibrate the dynamics of macroeconomic variables, energy sector outputs, and GHG emissions, especially in models that focus on the analysis of energy and climate policies (Foure et al. 2020; Faehn et al. 2020). They affect the development of different electricity technologies and impact their production pathways, portfolio of regional electricity technologies, and total \(CO_2\) emissions. Lastly, since results from CGE models are very sensitive to assumed values for trade elasticities as well as for elasticities of substitution (Mc Daniel and Balistreri 2003; Turner 2009, 2008), we also include two sets of elasticity parameters, namely, armel and esub_kle. Additional details of the DARTCLIM model are provided in the “Appendix”.
4.2 Iterative Approach
The main idea is to follow the same approach as in Sect. 3, using equation 3 to calibrate the selected DARTCLIM parameters. In general, Bayesian estimation could be performed using the original DARTCLIM model. However, since DARTCLIM is a recursive dynamic model, the optimization problem (3) is tedious to solve numerically even when powerful solvers are applied. Given that we want to calibrate over 1500 parameters in this application, the corresponding optimization problem requires a relatively large number of simulation runs, making it costly. Therefore, the efficiency gained by metamodelbased SO is important^{Footnote 8}.
The required number of simulations for a polynomial metamodel with interaction effects would also need a large number of simulations, and therefore, we only use local main effects. Firstorder polynomial metamodels are efficient and effective, provided they are “adequate” approximations (Kleijnen 2020). However, polynomial metamodels generally provide only local approximations, so a series of metamodels are estimated iteratively, optimizing goodness of fit. The developed iterative procedure follows the hillclimbing technique and wraps the steps from Algorithm 1 in a loop. Algorithm 2 gives an overview of the individual steps.
First, we draw a sample of parameters following a specific DOE. In particular, at each step k, we sample each parameter from an interval \([{\theta }^{lo}_{i,k},{\theta }^{up}_{i,k}]\), with the bounds being iteration specific^{Footnote 9}. Given that we are looking for local firstorder main effects, we generate a metamodel sample \({{\Theta }}^M\) where each parameter \({\theta }_i\) is sampled individually within its range, and all other parameters remain unchanged. The sampling strategy is based on the “one factor at a time” sampling method (Kleijnen 2015). In particular, we sample four points for each parameter \({\theta }_i\) with two points close to the current value and two points further away since some parameter values might result in infeasible simulations. The sample points \({{\Theta }}^M_k\), a \({\varvec{\theta }}*4 x {\varvec{\theta }}\) matrix, are calculated as follows:
In addition, a smaller validation sample \({{\Theta }}_k^V\) (500 parameter specifications) is generated using LHS (Sacks et al. 1989), where parameters \({\varvec{\theta }}_k\) are sampled simultaneously in the same ranges to test the predictive ability of the estimated metamodels. Computation of the individual runs of the metamodel and validation samples can easily be split into multiple parallel computations (Herlihy et al. 2020). The parallel computation is timesaving because a single simulation run with DARTCLIM takes from 15 to 60 minutes of computation time and we need to compute \(\approx 4 * 1500 + 500 = 6500\) different runs.
In the second step, we conduct simulation with the DARTCLIM model for each sampled parameter vector.
Next, we estimate and validate relevant metamodels \(M_k\) (see Sect. 2). In particular, we estimate a metamodel \(M_{k,j}\) for each output \({{Z}}_j\) and perform the validation.
In the fourth step, we conduct the Bayesian estimation of the model parameters \({\varvec{\theta }}^*_k\) based on the metamodels \(M_k\). Using these metamodels we can solve the slightly adapted problem 9 resulting in \({\varvec{\theta }}_k^*\):
where \(0 \equiv H({\varvec{\theta }})\) includes theoretical and empirical restrictions on the parameters including the iteration specific bounds (\({\varvec{\theta }}_k^{lo}, {\varvec{\theta }}_k^{up}\)).
In the fifth step, we use estimated parameter values \({\varvec{\theta }}^*_k\) to derive trend predictions of relevant output variables, \({\varvec{z}}^F_k \), from simulation runs of the original DARTCLIM model. If the prediction error (\({\varvec{z}^o}\) to \({\varvec{z}}^F_k\)) is below a critical threshold, then the algorithm stops, and the estimated parameters, \({\varvec{\theta }}^*_k\), are taken to calibrate DARTCLIM^{Footnote 10}. Otherwise, the process starts again at step 1, while we sample from the interval \([{\varvec{\theta }}^{lo}_{k+1}, {\varvec{\theta }}^{up}_k]\) defined around the estimated parameters \({\varvec{\theta }}^*_k\).
We repeatedly search for a local optimum in a restricted parameter space in this process. The subsequent parameter space for each parameter is set around the best solution found in the previous iteration. This iterative approach builds upon our described methodology (see Sect. 2), though it does not guarantee a global optimum since the iterative process might converge to a local optimum. However, the iterative process guarantees that goodness of fit predicting exogenous forecast variables is improved moving from one iteration to the next^{Footnote 11}.
The abovedescribed process is implemented in a combination of R (R Core Team 2020), GAMS (GAMS Development Corporation 2020) and Ansible scripts (Red Hat Software 2020). The Ansible scripts are used to automate the parallel computation in the cluster environment. The cluster environment uses the Slurm workload manager (Yoo et al. 2003). The DARTCLIM model and the Bayesian calibration are implemented in GAMS, while the sampling procedure and analysis are implemented in R (Wickham et al. 2019; Wickham 2016; Dupuy et al. 2015; Dirkse et al. 2020).
5 Results
In Sect. 5.1, we briefly evaluate the validation results by assessing the prediction performances of the metamodels. Subsequently, we assess the calibration results using a gof measure and compare the simulated trends with their forecast trends. Finally, we conclude the section with a short demonstration of the importance of baseline calibration on policy implications by simulating the impacts of fulfilling the emission reduction targets as committed under the Paris Agreement in the DARTCLIM model.
5.1 Validation
We use the metamodel sample to derive metamodels which are then used to make predictions of the validation sample. The iterative approach moves the model parameter space recurrently towards a parameter space, where the simulated trends are close to the forecast trends. We check whether the local firstorder polynomial metamodels for each iteration can approximate the underlying (I/O) relationships sufficiently well by assessing the prediction performance by means of the AER (see Equation 7):
where \({{Z}}^F_{k, j, u}\) and \({{Z}}^M_{k, j, u}\) denote the \(u^{th}\) simulated and predicted value for the \(j^{th}\) output variable in the \(k^{th}\) iteration. Moreover, we can see the evolution of the prediction performances across the iterations.
The results are shown in Fig. 3 for each output variable in each region and each iteration, which is denoted by the different colors (see Table 11 for detailed results).
The different scales on the yaxes provide a rough indication of the difficulty of using local firstorder polynomial metamodels to approximate the underlying (I/O) relationships of the DARTCLIM model, particularly in earlier iterations. Specifically, from Fig. 3 we observe that the (I/O) relationships for gdp, esolar and ewind are relatively easier to approximate than those for ffu, eother and semis as is evident from the lower AER values of the former three outputs relative to the latter three, especially in earlier iterations. A possible explanation for this could be that for ffu, eother and semis the underlying relationships are more complex than what the metamodels can capture, particularly when the model parameter space is large. Within each of these three variables, we also see regional differences in the magnitude of the prediction errors. For instance, the linear metamodels perform better at capturing the (I/O) relationships in eother in India than in eother in USA. This regional variation can be partly attributed to the linearity of the metamodels and, thus, their limitation in capturing the underlying relationship driven by factors like interaction effects across regions and sectors.
However, in the majority of the cases, the firstorder polynomials give satisfactory prediction performance and therefore, we still consider this assumption robust. Overall, from Fig. 3 it can be seen that for each output variable and across regions, the AER values tend to decrease with a higher number of iterations and finally reach relatively low values, thereby indicating fine prediction performances of the metamodels.
5.2 Baseline Calibration
The primary goal of our approach is to calibrate the baseline of dynamic models wherein the simulated trends from model outputs in the baseline scenario are close to their forecast trends. To this end, we use two measures to assess the calibration results. The first measure is again (gof) which is proportional to the (negative) likelihood part (\({{\mathcal {L}}({\varvec{y}^o}\mid {\varvec{\theta }})}\)) of the target function (see Equations 3 and 9). The second measure \(\kappa \) follows a more intuitive idea. We compare the simulated values \({\varvec{z}}_k^F\) (with \(F({\varvec{y}}, {\varvec{\theta }}_k^*) \equiv 0\)) with the corresponding forecasts \({\varvec{z}^o}\) and define an indicator function to count the number of output variables that fall in the predefined range:
with f a transformation function. In the simplest case f is just the identity function but it can also transform the CAGR into real values like GDP in \(US\$\).
To assess the results from our application, we first take a look at the gof measure. The results are given in Table 7 where “predicted” represents the gof computed by using \({\varvec{z}}_k^M = M_k({\varvec{z}}, {\varvec{\theta }}_k^*)\) while “simulated” represents the gof computed by using \({\varvec{z}}_k^F\).
The “predicted” gof values in the tot column measure the overall deviations of \({\varvec{z}}_k^M\) from \({\varvec{z}^o}\). The “predicted” gof values decreases with each iteration reflecting improvements in the prediction performance of metamodels. This pattern is also similar to the results seen in Sect. 5.1. Moreover, we can discern differences between the gof values for “predicted” and “simulated”. The differences are large in earlier iterations but subsequently reduce. Thus, the prediction performance of metamodels has a significant influence on the calibration process, and the improvement of the prediction performance leads to the refinement of the calibration point. After five iterations, we achieve an average prediction error below \(5\%\) of the calibrated baseline compared to the forecast trends.
Second, in order to have a more straightforward understanding of the results in the calibrated baseline scenario, we take a look at the results of the \(\kappa \) measure in Table 7. We transform the CAGR into real values for the year 2030 for each output variable in each region. This is important since small differences in CAGRs can lead to large differences in real values after accumulation over the years. From the last column in Table 7 we see that in the fifth iteration, 108 out of 120 simulated trends fall into the corresponding \(\pm 10\%\) ranges around the forecast trends, while in the first iteration, only 67 satisfy this criterion. Again, this demonstrates the advantages of the iterative approach and the improvement gains with additional iterations.
These results show that the iterative calibration method is capable of calibrating the baseline scenario of dynamic CGE model DARTCLIM in a structured and systematic way. The approach provides satisfactory calibration results that can be quantified, and it can be easily applied to a variety of simulation models. It should be noted that since the goal here is to illustrate the application of our method, we only performed five rounds of iteration. However, the error margins of the baseline calibration can be further reduced with further iterations.
5.3 Implications of an Improved Calibration Procedure on Policy Analysis
The main aim of parameter calibration is to use all available data that is informative regarding the specification of relevant model parameters. Our explanations demonstrate that using existing exogenous forecast data for parameter calibration is a rather complex and laborious process. Thus, one could finally ask whether it is worth the additional effort. This is an empirical question since parameters resulting from different calibration processes (adhoc static versus Bayesian dynamic methods) can imply significantly different economic responses to given policy shocks derived from the same CGE model.
An important response of economic systems to simulated climate policies is encapsulated in the induced change of marginal abatement costs. Hence, we use this variable to assess whether our dynamic calibration method induces significantly different economic responses than a standard calibration procedure. To this end, we impose in a simulated climate policy scenario regional unilateral \(CO_2\) reduction targets on all of the regions represented in DARTCLIM. The regional targets correspond to the conditional Nationally Determined Contributions (NDC)^{Footnote 12} that countries have pledged within the proceedings of the Paris Agreement under conditional emission reduction pledges. The emission reduction targets are defined as percentage reductions relative to the baseline.
We simulate the Paris Agreement policy scenario with DARTCLIM applying three different calibration methods implying three different parameter specifications of the model. First, we specify parameters applying our dynamic calibration method, labelled as full. Next, we use the standard static calibration method based on observed national SAMs in the base year with ad hoc assumed values for parameters driving future response. We refer to this specification method as static. Finally, we applied our dynamic calibration method again. However, in this case, we only matched three trends, namely regional GDP, total fossilfueled electricity, and \(CO_2\) emissions, and stopped the iterations after two rounds. In this case, we matched 48 out of 60 targets in a \(\pm 10\%\) range and 54 in a \(\pm 20\%\) range. This specification is labeled as subset.
As can be seen from Table 8 estimated carbon prices (marginal abatement costs) differ significantly across parameter specifications. Since the \(CO_2\) reduction targets are defined as percentage reductions relative to the baseline, the absolute emission reductions that need to be undertaken by regions differ across the three scenarios. Owing to the differences across baselines, the inherent economic tradeoffs for regions and sectors differ under the policy scenario. This difference gives to different estimates of abatement costs across regions. On average, percentage differences in national carbon prices equals \(\approx 21\%\) comparing standard parameter calibration (static) with our advanced dynamic calibration method (full), while on average, this difference amounts to \(\approx 16\%\) for the subset specification. Moreover, in some regions like PAS or BRA simulated carbon prices vary by \(\approx 50\%\) depending on the applied calibration procedures. We refrain from diving into further details about the results since the goal of this exercise is only to showcase that choices made in baseline calibration of CGE models are of crucial importance for further policy analysis.
6 Conclusion
Nowadays, evidencebased policies correspond to a standard approach of good governance. However, policy analysis is plagued by model uncertainty (Manski 2018). This holds especially true for CGEapplications as these models usually have weak econometric foundations. Accordingly, given that available data allowing a sophisticated econometric parameter specification remain rather limited, this paper develops an innovative, dynamic calibration method applying a Bayesian estimation framework. Bayesian estimation techniques are already successfully applied to estimate parameters of DSGE models. Like CGEs DSGE models correspond to sound microfounded and theoretically consistent model structures. Applying Bayesian estimation techniques DSGE combine these theoretical properties with good forecasting performance and hence have become an additional useful tool in the forecasting kit of central banks (Smets and Wouters 2003; Hashimzade and Thornton 2013). However, compared to CGE models DSGEs are still highly aggregated models, which assume highly aggregated economic agents, e.g., do not include heterogenous sectoral structures. Therefore, in contrast to DSGE models, Bayesian estimation techniques have rarely been applied to estimate parameters of a CGE model and even nowadays are still challenging. Technically, the Bayesian framework corresponds to an optimization problem. Given the complexity of most CGE models solution of the latter optimization problem cannot be solved with standard numerical solution algorithms. For example, the calibration of the DARTCLIM model includes over 1500 parameters to be specified.
In this context, this paper develops a metamodelbased simulation optimization (SO) approach. Given the fact that SO approaches generally require many simulation runs, and each run is quite expensive, the advantage of using metamodels is that they significantly reduce computation time, implying more efficient SO methods.
In particular, we first demonstrate how our approach can be applied using a toy model. Then, to make our approach accessible, especially to CGEmodellers, we explicitly supply all necessary code for replication. Second, we show the effectiveness of our approach by applying it to perform the calibration of a dynamic baseline of the DARTCLIM model based on forecasts of six central outputs. Our iterative calibration approach delivers convincing results. For example, for fitting values for about 1500 parameters using a total of 120 outputs, the prediction errors could be significantly reduced from iteration to iteration until the average prediction error led below an acceptable range of \(5\%\). Based on derived metamodels, Bayesian estimation of model parameters can be done quite effectively, requiring significant lower computational effort when compared to estimation procedures using the original CGE, i.e. it takes less than a minute. The derivation of the metamodels requires a significant computational effort, though, but that can be easily split across multiple computers or can be computed in a cluster environment.
Nevertheless, it is fair to conclude that compared to simple static calibration methods typically used in the literature, our advanced approach is definitely technically more demanding. Thus, the question arises: Is it worth the additional effort? In this regard, we demonstrate in this paper that economic response to climate policies encapsulated in regional emission prices derived from DARTCLIM differ by up to \(21\%\) depending on the applied calibration method. Hence, we conclude that investing additional resources into comprehensive parameter calibration is definitely worth the effort.
Furthermore, we illustrate that applying our Bayesian framework, the posterior distribution of the model can be used to calculate the complete distribution of the forecast and various risk measures proposed in the literature. In particular, the structural nature of the model allows computing the posterior distribution forecast conditional on a policy path. Moreover, it also allows examining the structural sources of the forecast errors and their policy implications. Technically, metamodeling also facilitates MetropolisHasting sampling from the posterior distribution. Thus, our method is not only a good method to explicitly reveal induced uncertainty of model like SSA, but also an appropriate procedure to reduce model uncertainty. Finally, it also allows for better reproducibility of policy research, as it replaces the adhoc (often trial and error) approaches with a theoretically founded and algorithmic approach.
Of course, many challenges still lie ahead. Here we particularly like to highlight three of those challenges that we consider interesting topics of future research. First, as Manski (2018) highlighted in his seminal paper, integrating fundamental model uncertainty is crucial for conducting comprehensive modelbased policy analysis. We certainly consider our Bayesian approach as a promising methodological starting point in this context. However, much more work is needed to both deepen and broaden the scope of existing CGEbased policy analysis. For example, when forecasting in a policy context, realistic assumptions about how practical politicians form their beliefs on how alternative policies impact relevant policy outcomes are crucial. In particular, political reality constraints on information processing, pervasive model uncertainty and the associated need for perpetual learning highlight the limits of the rational expectations assumption. Another topic for future research is to extend our calibration method to other complex models like interlinked ecologicaleconomic model frameworks. Furthermore, at a methodological level, it will be interesting to investigate other metamodeling techniques in the context of this approach, especially regarding computational intensity and prediction accuracy. In the iterative procedure used for large models, it appears worthwhile to investigate how to reuse information from previous iterations, like samples or even the estimated metamodels. The latter might be done by building a metamodel ensemble (BartzBeielstein and Zaefferer 2017; Friese et al. 2016), investigating if weighted linear combinations of metamodels improve the calibration procedure. The weights might be determined by a distance measure to the sample space of the respective metamodel. Given the advances in machine learning and automatic differentiation, in particular, it will also be interesting to investigate if those techniques can be used in our approach (Baydin et al. 2018).
Notes
The complete model of Hosoe et al. (2010, 2016) comprises of 62 behavioral parameters, \(4 \times 5 = 20\) input coefficients for the upper Leontief nest of the sectoral production functions, \(4 \times 2 = 8\) parameters for the four lower CobbDouglas (CD) production functions, \(4 \times 3 = 12\) parameters for the four Armington functions aggregating domestic supply and imports to domestic supply as well as \(4 \times 3 = 12\) parameters for the four CETfunctions transforming domestic production into domestic supply and exports for each commodity and finally 4 parameters for the CDdemand systems determining domestic demand of the representative domestic household. Finally, we have 6 parameters determining sectoral investments (\(\textit{iota}, \textit{zeta}\), and \(\textit{lambda(i)}\)). At this point we neglect exogenous variables, e.g. world market prices, tax rates and tariffs, etc...
Please note that generally, we could estimate metamodels for each time period or we could assume a specific growth function for every forecasted variable, \({{Y}}\), and estimate relevant growth rates as a metamodel.
Please note that to directly match variables in the GAMS and Rcode we kept the variable notation used in R also in GAMS, i.e. instead of defining in GAMS one variable \(\sigma ^2(J)\) where the set J includes all four production sectors, we defined four variables, \(\sigma ^2_j, J=\{AGR, LMN, HMN, SRV\}\).
This follows by construction as we generated the true forecast via simulation of the CGE based on the fixed true parameters, \({\varvec{\theta }}^{true}\).
Please note that assuming a \(1\%\) deviation on all target variables implies a value of \(\approx 141\).
For notational convenience we now denote \({\varvec{\theta }}_T\) with \({\varvec{\theta }}\), as we keep all other parameters fixed.
We set the critical threshold corresponding to an average prediction error of \(5\%\). Approximatively, this critical threshold reflects the relative information value of the forecasts visavis information encapsulated in the prior distribution of the calibrated DARTCLIM parameters.
It should be further noted that as long as the data on exogenous forecasts is itself noisy, parameter estimates implying perfect predictions are generally not desirable.
The method for calculating the emission reduction targets is based on Böhringer et al. (2021).
References
Aguiar, A., Narayanan, B., & McDougall, R. (2016). An overview of the GTAP 9 data base. Journal of Global Economic Analysis, 1, 181–208.
Arndt, C., Robinson, S., & Tarp, F. (2002). Parameter estimation for a computable general equilibrium model: A maximum entropy approach. Economic Modelling, 19(3), 375–398.
Barthelemy, J. F., & Haftka, R. T. (1993). Approximation concepts for optimum structural design—A review. Structural optimization, 5(3), 129–144.
BartzBeielstein, T., & Zaefferer, M. (2017). Modelbased methods for continuous and discrete global optimization. Applied Soft Computing, 55, 154–167. https://doi.org/10.1016/j.asoc.2017.01.039
Baydin, A.G., Pearlmutter, B.A., Radul, A.A., Siskind, J.M. (2018). Automatic differentiation in machine learning: a survey. Journal of Machine Learning Research 18(153), 1–43. http://jmlr.org/papers/v18/17468.html
Böhringer, C., Dijkstra, B., & Rosendahl, K. E. (2014). Sectoral and regional expansion of emissions trading. Resource and Energy Economics, 37, 201–225. https://doi.org/10.1016/j.reseneeco.2013.12.003
Böhringer, C., Peterson, S., Rutherford, T. F., Schneider, J., & Winkler, M. B. J. (2021). Climate policies after paris: Pledge, trade and recycle: Insights from the 36th energy modeling forum study (emf36). Energy Economics, 103
Breiman, L. (2001). Random forests. Machine Learning, 45(1), 5–32.
Burniaux, J. M., & Chateau, J. (2014). Greenhouse gases mitigation potential and economic efficiency of phasingout fossil fuel subsidies. International Economics, 140, 71–88. https://doi.org/10.1016/j.inteco.2014.05.002
Caliendo, L., & Parro, F. (2014). Estimates of the trade and welfare effects of NAFTA. The Review of Economic Studies, 82(1), 1–44. https://doi.org/10.1093/restud/rdu035
Chen, V. C., Tsui, K. L., Barton, R. R., & Meckesheimer, M. (2006). A review on design, modeling and applications of computer experiments. IIE Transactions, 38(4), 273–291.
Cressie, N.A.C. (1993). Statistics for Spatial Data. Wiley. 10.1002/9781119115151
Dey, S., Mukhopadhyay, T., & Adhikari, S. (2017). Metamodel based highfidelity stochastic analysis of composite laminates: A concise review with critical comparative assessment. Composite Structures, 171, 227–250.
Dirkse, S., Ferris, M., Jain, R. (2020). gdxrrw: An Interface Between ’GAMS’ and R. http://www.gams.com, r package version 1.0.6
Dixon, P. B., Jorgenson, D. W. (eds) (2013). Handbook of Computable General Equilibrium Modeling. NORTH HOLLAND
Dixon, P.B., Rimmer, M., Parmenter, B.R. (1997). The Australian textiles, clothing and footwear sector from 198687 to 201314: analysis using the Monash model. Monash University
Dupuy, D., Helbert, C., Franco, J. (2015). DiceDesign and DiceEval: Two R packages for design and analysis of computer experiments. Journal of Statistical Software 65(11):1–38, http://www.jstatsoft.org/v65/i11/
Faehn, T., Bachner, G., Beach, R.H., Chateau, J., Fujimori, S., Ghosh, M., HamdiCherif, M., Lanzi, E., Paltsev, S., Vandyck, T., et al. (2020). Capturing key energy and emission trends in cge models: Assessment of status and remaining challenges. Tech. rep.
Fan, Q., FisherVanden, K., & Klaiber, H. A. (2018). Climate change, migration, and regional economic impacts in the united states. Journal of the Association of Environmental and Resource Economists, 5(3), 643–671. https://doi.org/10.1086/697168
Fischer, C., Morgenstern, R.D. (2006). Carbon abatement costs: Why the wide range of estimates? The Energy Journal 27(2):73–86, http://www.jstor.org/stable/23297020
Forrester, A., Sobester, A., & Keane, A. (2008). Engineering design via surrogate modelling: a practical guide. John Wiley & Sons.
Foure, J., Aguiar, A., Bibas, R., Chateau, J., Fujimori, S., Lefevre, J., Leimbach, M., ReyLosSantos, L., & Valin, H. (2020). Macroeconomic drivers of baseline scenarios in dynamic cge models: review and guidelines proposal. Journal of Global Economic Analysis, 5(1), 28–62.
Friedman, J. H., et al. (1991). Multivariate adaptive regression splines. The annals of statistics, 19(1), 1–67.
Friese, M., BartzBeielstein, T., Emmerich, M. (2016). Building ensembles of surrogate models by optimal convex combination. Tech. rep., http://nbnresolving.de/urn:nbn:de:hbz:832cos43480
Fujimori, S., Kubota, I., Dai, H., Takahashi, K., Hasegawa, T., Liu, J. Y., Hijioka, Y., Masui, T., & Takimi, M. (2016). Will international emissions trading help achieve the objectives of the paris agreement? Environmental Research Letters, 11(10), 104001. https://doi.org/10.1088/17489326/11/10/104001
GAMS Development Corporation (2020) General Algebraic Modeling System (GAMS) Release 33.1. Washington, DC, USA, http://www.gams.com/
Gehlhar, M.J. (1994). Economic growth and trade in the pacific rim: An analysis of trade patterns. PhD thesis, Purdue University
Go, D. S., Lofgren, H., Ramos, F. M., & Robinson, S. (2016). Estimating parameters and structural change in CGE models using a bayesian crossentropy estimation approach. Economic Modelling, 52, 790–811. https://doi.org/10.1016/j.econmod.2015.10.017
Gomme, P., Lkhagvasuren, D. (2013). Calibration and simulation of DSGE models. In: Hashimzade N, Thornton MA (eds) Handbook of Research Methods and Applications in Empirical Macroeconomics, Chapters, Edward Elgar Publishing, chap 24, pp 575–592, https://ideas.repec.org/h/elg/eechap/14327_24.html
Gong, W., Duan, Q., Li, J., Wang, C., Di, Z., Dai, Y., Ye, A., & Miao, C. (2015). Multiobjective parameter optimization of common land model using adaptive surrogate modeling. Hydrology and Earth System Sciences, 19(5), 2409–2425.
Hansen, L. P., & Heckman, J. J. (1996). The empirical foundations of calibration. Journal of Economic Perspectives, 10(1), 87–104. https://doi.org/10.1257/jep.10.1.87
Hashimzade, N., & Thornton, M. A. (2013). Handbook of research methods and applications in empirical macroeconomics. Cheltenham, UK: Edward Elgar.
Heckelei, T., Mittelhammer, R. (2008). A bayesian alternative to generalized cross entropy solutions for underdetermined econometric models. Discussion Paper 2, Institute for Food and Resource Economics University of Bonn
Herlihy, M., Shavit, N., & Luchangco, V. (2020). The Art of Multiprocessor Programming. Oxford: Elsevier LTD.
Hertel, T., Hummels, D., Ivanic, M., & Keeney, R. (2007). How confident can we be of CGEbased assessments of Free Trade Agreements? Economic Modelling, 24, 611–635.
Hertel, T.W. (1997). Global Trade Analysis: Modeling and Applications. Cambridge Univ. Press, Cambridge [u.a.]
Hosoe, N., Gasawa, K., & Hashimoto, H. (2010). Textbook of Computable General Equilibrium Modelling. Palgrave Macmillan UK. https://doi.org/10.1057/9780230281653
Hosoe, N., Gasawa, K., Hashimoto, H. (2016). Textbook of Computable General Equilibrium Modeling: Programming and Simulations 2nd Edition. University of Tokyo Press, 10.1057/9780230281653
International Energy Agency. (2018). World Energy Outlook 2018. OECD
Jorgenson, D. (1984). Econometric Methods for Applied General Equilibrium Analysis, Cambridge University Press, Cambridge, pp 139–203. Growth 2, ch. 2, pp. 89155.
Kehoe, T. J., Polo, C., & Sancho, F. (1995). An evaluation of the performance of an applied general equilibrium model of the Spanish economy. Economic Theory, 6(1), 115–141. https://doi.org/10.1007/bf01213943
Kleijnen JP (2015) Design and analysis of simulation experiments. In: International Workshop on Simulation, Springer, pp 3–22
Kleijnen, J. P., & Sargent, R. G. (2000). A methodology for fitting and validating metamodels in simulation. European Journal of Operational Research, 120(1), 14–29.
Kleijnen JPC (2020) Simulation optimization through regression or kriging metamodels. In: HighPerformance SimulationBased Optimization, Springer International Publishing, pp 115–135, 10.1007/9783030187644_6
Kuik, O., Brander, L., & Tol, R. S. (2009). Marginal abatement costs of greenhouse gas emissions: A metaanalysis. Energy Policy, 37, 1395–1403.
Manski, C. F. (2018). Communicating uncertainty in policy analysis. Proceedings of the National Academy of Sciences, 116(16), 7634–7641. https://doi.org/10.1073/pnas.1722389115
Mareš, T., Janouchová, E., & Kučerová, A. (2016). Artificial neural networks in the calibration of nonlinear mechanical models. Advances in Engineering Software, 95, 68–81.
Mc Daniel, C. A., & Balistreri, E. J. (2003). A review of armington trade substitution elasticities. Economie internationale, 2, 301–313.
McKitrick, R. R. (1998). The econometric critique of Computable General Equilibrium modeling: The role of functional forms. Economic Modelling, 15, 543–573. https://doi.org/10.1016/S02649993(98)000285
de Melo, J. (1988). Computable general equilibrium models for trade policy analysis in developing countries: A survey. Journal of Policy Modeling, 10(4), 469–503. https://doi.org/10.1016/01618938(88)900178
MilczarekAndrzejewska, D., Zawalińska, K., & Czarnecki, A. (2018). Landuse conflicts and the common agricultural policy: Evidence from poland. Land Use Policy, 73, 423–433. https://doi.org/10.1016/j.landusepol.2018.02.016
Myers, R. H., Montgomery, D. C., & AndersonCook, C. M. (2016). Response surface methodology: process and product optimization using designed experiments. John Wiley & Sons.
Nong, D., & Simshauser, P. (2020). On energy and climate change policies: The impact of baseline projections. Applied Energy, 269, 115062. https://doi.org/10.1016/j.apenergy.2020.115062
OECD (2019) OECD Economic Outlook, Volume 2019 Issue 2. Paris, https://doi.org/https://doi.org/10.1787/9b89401ben, https://www.oecdilibrary.org/content/publication/9b89401b en
Olekseyuk, Z., & SchürenbergFrosch, H. (2016). Are Armington elasticities different across countries and sectors? A European study. Economic Modelling, 55, 328–342.
Paris, Q., Drogué, S., & Anania, G. (2011). Calibrating spatial models of trade. Economic Modelling, 28(6), 2509–2516. https://doi.org/10.1016/j.econmod.2011.07.008
Peters, J.C. (2016). The gtappower data base: Disaggregating the electricity sector in the gtap data base. Journal of global economic analysis, Global Trade Analysis Project (GTAP), Department of Agricultural Economics, Purdue University, West Lafayette, IN, https://www.gtap.agecon.purdue.edu/resources/ res_display.asp?RecordID=4619
Pothen, F., & Hübler, M. (2021). A forward calibration method for analyzing energy policy in new quantitative trade models. Energy Economics, 100,
R Core Team (2020) R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria, https://www.Rproject.org/
Rango, J., Schnorbus, T., Kwee, H., Beck, R., Kinoo, B., Arthozoul, S., Zhang, M. (2013). Comparison of different approaches for global modeling of combustion engines. Design of Experiments (DoE) in engine development pp 70–91
Razavi S, Tolson BA, Burn DH (2012) Review of surrogate modeling in water resources. Water Resources Research 48(7)
Red Hat Software (2020) Ansible. https://www.ansible.com/
Robinson, S., Cattaneo, A., & ElSaid, M. (2001). Updating and estimating a social accounting matrix using cross entropy methods. Economic Systems Research, 13(1), 47–64. https://doi.org/10.1080/09535310120026247
Ruben, R., & van Ruijven, A. (2001). Technical coefficients for bioeconomic farm household models: a metamodelling approach with applications for southern mali. Ecological Economics, 36(3), 427–441. https://doi.org/10.1016/s09218009(00)002408
Sacks J, Welch WJ, Mitchell TJ, Wynn HP (1989) Design and analysis of computer experiments. Statistical science pp 409–423
Sánchez, M. V. (2004). Rising inequality and falling poverty in Costa Rica’s agriculture during trade reform: a macromicro general equilibrium analysis. Shaker Publishing, Maastricht, chap Chapter, 7, 189–226.
Sanchez, S.M. (2006). Work smarter, not harder: Guidelines for designing simulation experiments. In: Proceedings of the 2006 Winter Simulation Conference, IEEE, 10.1109/wsc.2006.323037
Shan, S., Wang, G.G. (2010). Metamodeling for high dimensional simulationbased design problems. Journal of Mechanical Design 132(5), 10.1115/1.4001597
Shoven, J. B., & Whalley, J. (1984). Applied generalequilibrium models of taxation and international trade: An introduction and survey. Journal of Economic Literature, 22(3), 1007–1051. https://doi.org/10.2307/2725306
Shoven, J.B., Whalley, J. (1992). Applying General Equilibrium. Cambridge University Press, https://EconPapers.repec.org/RePEc:cup:cbooks: 9780521266550
Simpson TW, Peplinski J, Koch PN, Allen JK (1997) On the use of statistics in design and the implications for deterministic computer experiments. Design Theory and MethodologyDTM’97 pp 14–17
Simpson, T. W., Lin, D. K., & Chen, W. (2001). Sampling strategies for computer experiments: design and analysis. International Journal of Reliability and Applications, 2(3), 209–240.
Smets, F., & Wouters, R. (2003). An estimated dynamic stochastic general equilibrium model of the euro area. Journal of the European Economic Association, 1(5), 1123–1175. https://doi.org/10.1162/154247603770383415
Smith, M. (1993). Neural networks for statistical modeling. Thomson Learning
SobieszczanskiSobieski, J., & Haftka, R. T. (1997). Multidisciplinary aerospace design optimization: survey of recent developments. Structural Optimization, 14(1), 1–23. https://doi.org/10.1007/bf01197554
Stifel, D. C., & Thorbecke, E. (2003). A dualdual CGE model of an archetype african economy: trade reform, migration and poverty. Journal of Policy Modeling, 25(3), 207–235. https://doi.org/10.1016/s01618938(03)000140
Tarp, F., Arndt, C., Jensen, H.T., Robinson, S., Heltberg, R. (2002). Facing the development challenge in Mozambique: an economywide perspective. resreport 126, International Food Policy Research Institute (IFPRI)
Taylor, J. E., YunezNaude, A., & Dyer, G. (1999). Agricultural price policy, employment, and migration in a diversified rural economy: A villagetown CGE analysis from mexico. American Journal of Agricultural Economics, 81(3), 653–662. https://doi.org/10.2307/1244030
Thube, S. D., Delzeit, R., & Henning, C. H. (2022). Economic gains from global cooperation in fulfilling climate pledges. Energy Policy, 160,. https://doi.org/10.1016/j.enpol.2021.112673
Turner, K. (2008). A computable general equilibrium analysis of the relative price sensitivity required to induce rebound effects in response to an improvement in energy efficiency in the uk economy. SIRE Discussion Papers 200820, Scottish Institute for Research in Economics (SIRE), https://ideas.repec.org/p/edn/sirdps/33.html
Turner, K. (2009). Negative rebound and disinvestment effects in response to an improvement in energy efficiency in the UK economy. Energy Economics, 31(5), 648–666. https://doi.org/10.1016/j.eneco.2009.01.008
Vapnik, V. (2013). The nature of statistical learning theory. Springer science & business media
VillaVialaneix, N., Follador, M., Ratto, M., & Leip, A. (2012). A comparison of eight metamodeling techniques for the simulation of n2o fluxes and n leaching from corn crops. Environmental Modelling & Software, 34, 51–66.
Wang, H., Tang, L., & Li, G. (2011). Adaptive MLSHDMR metamodeling techniques for high dimensional problems. Expert Systems with Applications. https://doi.org/10.1016/j.eswa.2011.04.220
Wickham, H. (2016). ggplot2: Elegant Graphics for Data Analysis. SpringerVerlag New York, https://ggplot2.tidyverse.org
Wickham H, Averick M, Bryan J, Chang W, McGowan L, François R, Grolemund G, Hayes A, Henry L, Hester J, Kuhn M, Pedersen T, Miller E, Bache S, Müller K, Ooms J, Robinson D, Seidel D, Spinu V, Takahashi K, Vaughan D, Wilke C, Woo K, Yutani H (2019) Welcome to the tidyverse. Journal of Open Source Software 4(43):1686, 10.21105/joss.01686
Winkler, M. B. J., Peterson, S., & Thube, S. (2021). Gains associated with linking the eu and chinese ets under different assumptions on restrictions, allowance endowments, and international trade. Energy Economics, 104,
Yildizoglu, M., Salle, I., et al. (2012). Efficient sampling and metamodeling for computational economic models. Groupe de Recherche en Economie Théorique et Appliquée: Tech. rep.
Yoo, A.B., Jette, M.A., Grondona, M. (2003). SLURM: Simple linux utility for resource management. In: Job Scheduling Strategies for Parallel Processing, Springer Berlin Heidelberg, pp 44–60, 10.1007/10968987_3
Acknowledgements
We thank anonymous reviewers for critically reading the manuscript and suggesting substantial improvements.
Funding
Open Access funding enabled and organized by Projekt DEAL. The research leading to these results has received funding from the Federal Ministry of Education and Research (BMBF) under grant agreement no. 031B0231A/B (MOD_GW_NW).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix
A Dart
DARTCLIM is a recurisve multiregion, multisector CGE model. The model version used in the application is based on the GTAP9 database. The primary goal of this model version is analysis of climate and energy policies. Representation of sectoral details affects the policy impact anaylsis in CGE model(Fischer and Morgenstern 2006; Kuik et al. 2009). Therefore, in order to have a rich representation of the energy sectors, data about electricity production by technology is additionally included from the GTAP  Power database (Peters 2016). In the GTAP9  Power database the supply of some of these technologies is differentiated by baseload and peakload. However, in DARTCLIM we have aggregated the baseload and peakload sectors to a single homogeneous sector.
Overall DARTCLIM has eight different technologies that produce electricity: solar PV, wind, nuclear, hydro, coal, gas, oil, and other renewable technologies. We assume that the development of nuclear and hydroelectric technologies is significantly determined by political willpower and/or geographical framework conditions, and less by price effects in electricity markets. Accordingly, the production pathways of these two technologies is defined exogenously in the model based on data from International Energy Agency (2018). The production activites are defined as CES functions in the model. The details about the nesting structures for different sectors are available in (Thube et al. 2022).
.
B Results
See Table 11
.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ziesmer, J., Jin, D., Thube, S.D. et al. A Dynamic Baseline Calibration Procedure for CGE models. Comput Econ 61, 1331–1368 (2023). https://doi.org/10.1007/s10614022102484
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10614022102484