1 Introduction

Uncertainty in climate science has drawn increasing attention in recent years (e.g., Parker 2006, 2010, 2011, 2013; Stainforth et al. 2007; Knutti 2008; Frigg et al. 2013, 2014; Parker and Risbey 2015). The topic is important epistemically and politically: epistemically, because scientists have only limited abilities to validate and confirm the output of climate modelsFootnote 1; and politically, because policymakers have to take into account the current knowledge concerning the climate and its uncertainty.Footnote 2

There are different types of uncertainty in climate science (Knutti 2008; Parker 2010). This paper focuses on structural uncertainty, that is, uncertainty about the mathematical structure of a climate model.Footnote 3 My goal is to help develop an epistemological account of structural uncertainty that pays close attention to the details of climate modeling. As I show below, the detailed look at model building is needed if an epistemology of structural uncertainty is to be useful to scientists and policymakers.

The paper has two parts. The first part (Sects. 2, 3) motivates the need for an epistemological account of structural uncertainty that is informed of the details of model building. I begin by reviewing some of the existing accounts of structural uncertainty to illustrate how they have identified the sources of structural uncertainty. But a useful account of structural uncertainty should do more than identifying the sources of this uncertainty. To clarify what a useful account must do, I present three desiderata and argue that in order for an account of structural uncertainty to meet the desiderata, it needs to be informed of the details of model building. Accordingly, the second part of the paper (Sect. 4) develops one important detail for an epistemological account of structural uncertainty: the assumptions behind the choices of parameterization schemes, their conceptual justificatory role and their empirical adequacy.Footnote 4

Climate modelers introduce parameterizations in their models to code for the effect of physical processes not explicitly represented in the models on those processes that are explicitly represented. Although parameterizations may seem like something to be removed whenever possible, I will argue that parameterizations can have epistemic merits: they can improve our understanding of the interaction of coherent structures observed in the climate system across spatiotemporal scales (e.g. hurricanes, clouds, and squall lines).Footnote 5 Parameterizations can do this because they eliminate some of the details about the behavior of smaller scale physics that may be irrelevant for the modeling of these coherent structures at a particular scale. Moreover, I argue that the epistemic utility of parameterizations depends on whether they are properly justified theoretically and empirically. If the scale separation assumption can be empirically verified, structural uncertainty about a given parameterization scheme will decrease. At the end of the second part, I briefly show how this detailed look at model building helps an epistemological account of structural uncertainty to meet the desiderata presented in the first part.

2 Current accounts of structural uncertainty

The leading accounts of structural uncertainty are given by Frigg et al. (2014) and Parker (2006, 2010, 2011). Let us look at them briefly.

2.1 Frigg et al

Frigg et al. (2014) present an account of structural uncertainty in the context of nonlinear dynamical models.Footnote 6 In their view:

A model has [structural uncertainty] if the model dynamics differ from the dynamics in the target system…. [I]f a nonlinear model has only the slightest [structural uncertainty], then its ability to generate decision-relevant probability predictions is compromised (Frigg et al. 2014, p. 32).Footnote 7

This account implies that for non-linear systems, small differences in choice of model structure can compromise the ability to generate decision-relevant predictions. Frigg et al. call this effect of structural uncertainty the “hawkmoth effect” (Frigg et al. 2014, p. 39, after Thompson 2013), which is similar to sensitive dependence on initial conditions for chaotic systems. As Frigg et al. say, for both cases, it does not matter how close the model structure or the initial data are to the structure or data that will allow the model to make accurate predictions, because any small deviation in model structure or data can lead to misleading predictions.

The hawkmoth effect suggests that climate models are not reliable for policymaking. But can we manage the hawkmoth effect or determine its severity? Frigg et al. are pessimistic about being able to obtain a non-arbitrary measure on a class of models that would allow scientists to determine which models are or are not reliable. Thus, they argue that assigning probabilities to model-based predictions is a misleading way to represent predictions (Frigg et al. 2014, p. 57).

2.2 Parker

Parker’s account of structural uncertainty identifies sources of this uncertainty in more detail. Paraphrasing her account, structural uncertainty arises when:

  • Physical processes of interest are not described by well-established theories.

  • The representation of physical processes involves simplifications and idealizations, because of theoretical or pragmatic constraints.

  • Processes need to be parametrized, and there is no best way to parametrize a process. (Parker 2010, pp. 264–265)

Here, as in Frigg et al.’s account, structural uncertainty is an unavoidable aspect of climate modeling, because the use of simplifications, idealizations and parameterizations is unavoidable in climate modeling.Footnote 8

2.3 Toward an epistemology of structural uncertainty

The accounts of structural uncertainty given above identify its sources. In particular, Parker identifies sources within the practice of climate modeling, and as Frigg et al. argue, structural uncertainty poses serious challenges to the way climate models are currently used for policymaking. But an epistemology of structural uncertainty must do more than explaining why this type of uncertainty is hard to assess (e.g., Stainforth et al. 2007; Knutti 2008; Parker 2010; Frigg et al. 2014). Both scientists and policymakers need an epistemological account of structural uncertainty that facilitates communication about the uncertainty of model-based statements within and across the different disciplines. By model-based statements, I mean predictive or conditional statements about the future of the earth’s climate as well as statements about the uncertainty associated with these predictions or conditional statements.Footnote 9 As Stainforth et al. say:

A two-way communication between climate scientists and users of climate science is… of fundamental importance. Only by understanding the needs of different [policy] sectors can the science be usefully directed and communicated. Only by understanding the conditions, assumptions and uncertainties of model-based statements about future climate can decision makers evaluate the relevance of the information and make informed, if subjective, assessments of risk. (Stainforth et al. 2007, p. 2165; my emphasis).Footnote 10

In other words, an effective communication between scientists and policymakers requires an understanding of policymakers’ needs and an understanding of the uncertainties that affect model-based statements and the relevance of the uncertainties to policy. An epistemology of structural uncertainty, in my view, should aim to contribute to the latter understanding.Footnote 11

To achieve this aim, what exactly is needed for an epistemological account of structural uncertainty? In the next section, I present specific desiderata for this account.

3 Desiderata

As we saw, Frigg et al.’s and Parker’s accounts of structural uncertainty identify sources of this uncertainty. This is the first thing an epistemological account of structural uncertainty should do. Beyond that, in my view, the account should explain epistemic reliability of climate models and show how structural uncertainty can be relevant to specific policy questions. Let me go over these desiderata in turn.

3.1 Desideratum 1: Identifying sources of uncertainty

An adequate epistemological account of structural uncertainty should identify sources of uncertainty in the model and the model building practice. That is, it should say what makes climate models structurally uncertain. Philosophers have so far focused on this desideratum: they have identified aspects of the model construction and justification process that are likely to introduce uncertainty. To do so, their strategy has been to show that certain model assumptions—especially those that introduce simplifications and idealizations—generate structural uncertainty.

Identifying the sources of uncertainty should be something an account of structural uncertainty should do because doing so helps explain why a model is not likely to give accurate predictions, that is, why a model is not epistemically reliable.

3.2 Desideratum 2: Explaining epistemic reliability

But an adequate account of structural uncertainty should also explain why a model or its component is epistemically reliable.Footnote 12 By epistemic reliability, I have in mind how likely a model gives, or how likely its component helps give, accurate model-based statements, such as predictions about the future climate.Footnote 13 From experience we can learn that a model is epistemically reliable, just as I can learn that my calculator is epistemically reliable. The second desideratum here asks us to explain why a model or its component is reliable.

This desideratum becomes particularly important in the context of climate change, where scientists are asked to predict states of the climate that have never been observed before (Stainforth et al. 2007). So, this desideratum is important for an epistemological account of structural uncertainty because understanding why a model or its component is reliable helps us to determine the reliability of the model or its component in domains other than those we have explored.Footnote 14

Consider, for example, the damped harmonic oscillator model. It is epistemically reliable for predicting the movement of a weight attached to a spring in a viscous medium when the motion is sufficiently slow.Footnote 15 And we know why: the model is constructed by using Newton’s second law of motion and Hooke’s law—which apply to the system of interest—and a damping coefficient (an empirical parameter) that describes the viscous drag (or friction) exerted on the weight. This understanding helps us to determine whether the model will be reliable in a new domain.

By contrast, our understanding of the reliability of weather prediction models relies largely on their track record. These models are used to forecast local weather and are considered reliable mostly for predictions of 1–7 days.Footnote 16 While these models are based on physical theory, they are heavily calibrated to work in the regions in which they are applied. Moreover, they are analyzed in ensembles, which are then empirically corrected for the kind of atmospheric conditions that arise locally.Footnote 17 For example, we cannot reliably take a weather forecasting model calibrated to work for a coastal region in India and use it for a coastal region in the Caribbean without some substantial modification, tuning and validation for the conditions of the Caribbean. It is therefore not always clear why any one of these models is reliable.

The epistemic situation is worse for the case of climate models, since in that case we cannot evaluate a model based on its predictive success.Footnote 18 In these cases, explanations of reliability must rely on the details of model building: why are the model components an adequate representation of the target? How do modeling choices affect the applicability of a model? Explaining epistemic reliability aims at clarifying these questions.

3.3 Desideratum 3: Showing policy relevance

An adequate account of structural uncertainty should show how the severity of structural uncertainty can vary depending on questions asked by the policymakers. As Stainforth et al. (2007) suggest, in analyzing climate models, climate scientists should take into account the questions of the policymakers and the information they require. By doing so, scientists can hope to effectively communicate the limits of their models and clarify the extent to which certain analyses of structural uncertainty are tied to the applicability of models. Now, policymakers might expect climate models to be epistemically reliable with respect to particular phenomena at certain spatial and/or temporal scales. And scientists might understand why models are not likely to be reliable in the way expected by policymakers because they identify many sources of structural uncertainty. Scientists might also be able to point out the range of predictions that can be made reliably with the models because they understand why the models are reliable. The third desideratum therefore asks an adequate account of structural uncertainty to show how the policy-relevant importance of structural uncertainty varies with the questions and expectations of policymakers.

One might wonder why an epistemological account of structural uncertainty should show the relevance of structural uncertainty to policymakers. Why should an epistemology of structural uncertainty address policy? My answer is that if scientists are to produce information that can be acted upon by society, policymakers need to know the reliability of climate models with respect to their own questions. Suppose we have an account that meets the first and second desiderata, and we say why a model is unreliable and why it is, or should be, reliable in this or that domain. But we may still fail to be answering the policymaker’s epistemological problem. Thus, I suggest we include a separate desideratum—the third one here—specifically asking an epistemological account of structural uncertainty to explicitly take into consideration how structural uncertainty changes according to the questions and expectations of policymakers.

4 Structural uncertainty through the lens of model building

Thus far I have argued for the need to develop an epistemological account of structural uncertainty that meets the three desiderata given above. In this section, I develop a part of the account that meets these desiderata.Footnote 19 To meet all these desiderata, we have to look at the details of model building. Because to identify sources of structural uncertainty, to explain why a given climate model is reliable in a given domain, or to show how structural uncertainty of a model is or is not relevant to policy, we have to understand how climate model building works.

The account of structural uncertainty I develop below makes a number of claims:

  1. (1)

    Parameterizations can help us understand the across-scale interactions of different components of the climate (Sect. 4.1).

  2. (2)

    This epistemic merit of parameterizations depends on the availability of theoretical and empirical justification for them (Sect. 4.1).

  3. (3)

    Scale separation assumptions play an important role in the development of certain parameterization schemes (Sect. 4.2).

  4. (4)

    In modeling, the scale separation assumptions provide theoretical justification for parameterizations in the form of implicit or explicit equilibrium arguments (Sect. 4.2).

  5. (5)

    In modeling, parameterizations are partially justified empirically by evidence for the scale separation assumptions (Sect. 4.3).

  6. (6)

    If parameterizations are not justified theoretically or empirically, they have more structural uncertainty (Sect. 4.4).

  7. (7)

    But if parameterizations are justified theoretically and empirically, they provide a partial explanation of why a climate model is epistemically reliable with respect to phenomena and their interactions at certain scales (Sect. 4.4)

  8. (8)

    Analyzing justifications for parameterizations helps address the policymaker’s epistemological question (Sect. 4.5)

As can be seen, my account of structural uncertainty centers on parameterizations and scale separation assumptions, and there are two reasons for this focus. First, as noted above in relation to Parker’s account of structural uncertainty, simplifications are often a source of structural uncertainty. Parameterizations are a kind of simplification in climate models and are a major source of structural uncertainty. My account aims to analyze the extent to which parameterizations introduce structural uncertainty. Second, as I show below, climate modelers sometimes use equilibrium arguments, which imply the scale separation assumption, to provide theoretical justification for parameterization schemes, and they also try to provide empirical support for the scale separation assumptions. This justificatory practice is different from introducing parameterization schemes ad-hoc. Ad hoc parameterizations are a severe source of structural uncertainty, but theoretically and/or empirically justified parameterizations can reduce structural uncertainty at the scales at which they are justified. Thus, my account aims to articulate the kinds of justification that are available for parameterizations.

Before going into the details of how scale separation is used in modeling, it is worth clarifying the relation between spatiotemporal scales and phenomena. Bogen and Woodward (1988) defined phenomena as coherent, repeatable patterns in nature. I interpret phenomena in this sense as target systems in climate modeling, and they occur at many different spatiotemporal scales (Emanuel 1986). That is, phenomena depend for their existence on spatiotemporal scales.Footnote 20 Since the climate is complex and climatic phenomena occur across many scales, it is difficult to identify phenomena at different scales and model their interaction across scales. This is a main goal of climate science in so far as it aims to produce policy-relevant knowledge. For example, scientists are interested in knowing how the increase in global average temperature (a large-scale phenomenon) will affect smaller scale phenomena like hurricane frequency and intensity in the Caribbean (Emanuel 1999, 2005).

4.1 The need for parameterizations

Modelers introduce parameterizations because of pragmatic and epistemic constraints (McFarlane 2011). When understood as a constraint, a parameterization is an element of the model that should be removed as these constraints are overcome. However, increasing the resolution of a model, i.e. decreasing the number of parametrized processes, does not always reduce structural uncertainty (Van der Sluijs et al. 1998; McWilliams 2007; McFarlane 2011; Knutti and Sedláček 2013). So, are there epistemic merits to parameterizations that are relevant to an account of structural uncertainty? Akio Arakawa, who developed one of the most widely used cloud parameterization schemes, suggests that parameterizations do have epistemic merits:

Even under a hypothetical situation in which we have a model that resolves all scales, it alone does not automatically give us an understanding of scale interactions. Understanding inevitably requires simplifications, including various levels of “parameterizations”…. Parameterizations thus have their own scientific merits (Arakawa 2004, p. 2496).

According to Arakawa, parameterizations can be simplifications that aid the understanding of interactions across scales. Understanding how phenomena interact across scales can reduce structural uncertainty as it contributes to the theoretical understanding that increases the epistemic reliability of a model. And across-scale interactions are a recognized source of structural uncertainty (Slingo et al. 2003). In what follows, I will provide an argument for Arakawa’s claim.Footnote 21

Parameterizations are ways to account for the effect of unresolved processes on the resolved ones. So, in general, to have any epistemic merit they need to account for this effect for the right reasons. One such reason would be a physically principled argument for the use of a given parameterization scheme. Another reason would be the empirical evidence for the assumptions used in the physical argument. The next subsections are going to provide an example of how climate scientists have constructed these arguments and how they have appealed to empirical evidence.

4.2 Scale separation and parameterizations

One of the main strategies used to develop parameterization schemes is to implicitly or explicitly assume that there is some kind of physical or statistical equilibrium between the resolved scale and the unresolved processes (see, e.g., Williams 2005, p. 2935; McFarlane 2011, p. 491). This equilibrium assumption is a scale separation assumption: at equilibrium, the smaller scale processes occur on scales that are so fast (or so small) that they do not matter for the larger scale description. In fact, parameterization schemes often appeal to scale separation in order to distinguish between large scale and small scale phenomena (Plant and Craig 2008, p. 89; Arakawa 2004). For example, the scale separation assumption is used to justify the parameters appearing in equations that describe Newtonian fluids, such as viscosity.Footnote 22 The assumption that there is a separation of scale is introduced independently of whether any such separation can be detected in observations (Plant and Craig 2008, p. 90).Footnote 23

An illustrative example from climate science of a parameterization is the Arakawa-Schubert scheme for the parameterization of cumulus convection (Arakawa and Schubert 1974; Arakawa 2004). Cumulus convection is a physical process that is responsible for the persistence of cumulus clouds.Footnote 24 The Arakawa-Schubert parameterization attempts to code for the collective effect of smaller scale phenomena on the persistence of a group of clouds in a slowly varying environment, and the effect of these clouds on the environment. In particular, this parameterization scheme does not explicitly represent the processes that occur internally to each cloud, such as the conversion of humidity into rain droplets, but it explicitly represents the (vertical) mass flux of moist air into and out of the collective group of clouds (Arakawa 2004, p. 2498).

Now, this parameterization is justified in terms of a quasi-equilibrium of the mass flux with its environment: the environment creates an instability that is favorable for clouds to entrain (influx) moisture from the environment, and the clouds respond via a stabilization during which the clouds detrain (outflow) moisture to the environment. This quasi-equilibrium condition is warranted if the microphysical processes internal to the cloud occur at time scales that are sufficiently small to be considered to change instantaneously with its large scale environment (Arakawa 2004, p. 2505). However, it is unclear whether any such scale separation can be observed (Arakawa 2004, p. 2496).

Independently of whether there is any empirical verification of the use of this equilibrium assumption, we can ask what the physical insight of using this assumption might be. For sufficiently large scales, the effect of a group of clouds can be represented in terms of their average contribution to the mass flux of humidity. At these particular scales, the clouds can be considered to be in quasi-equilibrium with their environment. Thus, for the purpose of modeling the large scale environment, the Arakawa-Schubert parameterization explains the coherence and stability of a group of cumulus clouds in terms of the quasi-equilibrium of mass-flux.Footnote 25

It is important to note that this or any parameterization scheme will only work under specific atmospheric conditions. As atmospheric conditions change, due to anthropogenic greenhouse gas emissions or otherwise, moisture availability, chemical composition of the atmosphere and temperature conditions change too. These changes will affect the applicability of this parameterization scheme and the validity of its justification. So, the applicability of parameterizations changes as the climate changes, and this can be a source of structural uncertainty.Footnote 26 Nevertheless, highlighting the conditions under which modelling assumptions is valid is a starting point for an account that addresses the applicability of models.

4.3 Empirical justification of scale separation assumptions

We have seen so far that the scale separation assumption is one of the key ingredients in justifying parameterization schemes. Thus, whether a parameterization scheme introduces structural uncertainty can depend on whether the scale separation assumption used to justify the parameterization is itself warranted.

A scale separation assumption is clearly warranted when such a relevant scale separation is observed in the time series data for the parameters of interest. In such cases, the parameterization justified by the scale separation assumption is not likely to introduce substantial structural uncertainty.Footnote 27 However, as noted above, such a scale separation is rarely observable in climate data, and not all parameterization schemes are formulated with parameters that are easily observable.

Another empirical interpretation of scale separation is in statistical terms. The Arakawa-Schubert parameterization described above, for example, requires that the scheme describes a sufficiently large cloud ensemble, such that individual fluctuations in the parameters that describe the microphysics of clouds are negligible for the quasi-equilibrium assumption to hold (Arakawa and Chen 1987, p. 117). So, this scheme will be applicable only if there is a sufficiently large number of cumulus clouds in the target system.Footnote 28

Both the theoretical and the empirical aspects of the parameterization schemes can help explore structural uncertainty. For example, when the scale separation cannot be clearly detected in the data, but the theoretical justification is well received, there are various strategies that are taken to explain the lack of scale separation.

One such strategy is to add a stochastic component to the parameterization scheme (see Bony et al. 2015). This is the main approach that has been taken for the case of the Arakawa-Schubert parameterization, where scale separation is contested.Footnote 29 The idea behind this strategy is that while the physical explanation of the stability of the system (e.g. quasi-equilibrium) is appropriate, the scale separation is only loosely observable (Plant and Craig 2008, pp. 87–88). For these cases, then, it is suggested to adopt a stochastic approach (Plant and Craig 2008; Berner et al. 2017), and the uncertainty related to this kind of parameterization can be explored by having a large enough ensemble of models that captures the variability of the system with respect to this parameterization (Plant and Craig 2008).

Another strategy is to develop different parameterizations altogether, that do not rest on assumptions of equilibrium. Giving up equilibrium assumptions may mean that different kinds of mechanisms may be responsible for the observed stability (e.g. Yano et al. 2012), or that no principled parameterization can be obtained. In the former case, the effectiveness of different parameterization schemes in predicting the target of interest—how an ensemble of clouds behaves in a large scale environment—at the appropriate scale can be explored to account for structural uncertainty. In the latter case, the problem collapses to the case of ad hoc parameterizations.

While not explicitly stated in the context parameterizations, this kind of theoretical understanding has been recognized as being important for systematically exploring representational issues in climate models: Bony et al. (2006, p. 3447), for example, suggest that understanding physical mechanisms behind empirical estimates of parameters in the climate (in their case, climate feedbacks), can help understand why different models differ in their estimates, how reliable different models can be for these estimates, and guide the development of strategies to compare model output with observational data. In other words, this kind of reasoning can guide a systematic exploration of one important aspect of structural uncertainty. To summarize, the epistemic merits of parameterizations are that in some cases, if well justified, they can help the development of understanding of the interactions of different components of the climate across scales. And, as a minimum, systematically exploring justificatory strategies of parameterizations can aid the analysis of the structural uncertainty introduced by parameterizations.

4.4 Sources of structural uncertainty and partial explanation of reliability

Studying structural uncertainty from the perspective of model building allows the philosopher and the scientist to shed light on the assumptions that generate structural uncertainty. My analysis shows that both theoretical justifications and empirical considerations are used for the development of certain parameterization schemes. Structural uncertainty stemming from parameterizations, then, is a function of how well justified and how empirically accurate these assumptions are. We can now ask, does this approach meet the desiderata described above?

The first desideratum is that an account of structural uncertainty should reveal sources of uncertainty in the model. Frigg et al. (2014) and Parker (2006, 2010, 2011) already do so: nonlinearity puts limits to probabilistic model predictions, and various aspects of modeling strategies, theoretical or pragmatic, introduce structural uncertainty. Focusing on the details of modeling strategies such as the use of parameterizations can give a more detailed account of the severity of structural uncertainty. And while structural uncertainty may indeed be an irreducible aspect of models (McWilliams 2007), it can be explored systematically.Footnote 30

The second desideratum is that an account of structural uncertainty should indicate when and why a model is epistemically reliable, i.e. an account should articulate the conditions under which a model is likely to predict the behavior of its target accurately. Focusing on how parameterizations are introduced in models and how they are justified can shed light on this desideratum: when parameterizations are justified theoretically and verified empirically, there is good reason to believe that, for those scales at which the parameterization is relevant, the model is likely to predict the behavior of its target reliably.

The content of a scale separation assumption also indicates the domains of inapplicability of a particular model: the parameters that are introduced in a model will be a constraint on the kind of processes that are predicted by the model and the predictions that cannot be made by the model. For example, the Arakawa-Schubert parameterization can only be applied when the target has a sufficiently large number of cumulus clouds to warrant the use of the statistical equilibrium assumption to separate scales of motion.

Another implication of this analysis is that, since the assumption effectively ignores details at smaller scales, in order to obtain reliable downscaled models, scientists will need to identify what phenomena are relevant at various smaller scales and how they interact with phenomena at other scales. A downscaling process that does not take theoretical constraints of the scale separation assumption into consideration will end up being unreliable.Footnote 31

4.5 Policy relevance

The third desideratum concerns way structural uncertainty depends on the question that is asked of models: for what parameters and spatiotemporal scales is it possible to explore structural uncertainty systematically? This question is important because policy relevant knowledge also needs to accurately represent the conditions under which these knowledge claims are made, and the uncertainties that are tied to them.

While parameterizations are only one component of models that can contribute to model uncertainty, it does suggest that structural uncertainty may vary depending on the scale and parameter of interest, and scientists should make this explicit when giving estimates of uncertainty. However, more interesting policy-relevant questions arise when structural uncertainty cannot be thoroughly assessed for those predictions required by policymakers: for example, when alternative parameterization schemes are not investigated when a particular one is not well justified.Footnote 32 In these cases, is scientific information decision relevant? There are serious issues, in fact, with climate predictions for which the assessment and representation of uncertainty has not been taken into consideration thoroughly (Frigg et al. 2013, 2014, 2015).

5 Conclusion

In this paper, I have argued that philosophers have identified an important problem in climate science; namely, how best to assess and communicate the nature of structural uncertainty in climate modeling. However, their characterization of structural uncertainty is not sufficiently well-explicated. I have argued that a useful account of structural uncertainty should meet desiderata that are driven by scientific, philosophical and policy motivations. By focusing on the context of model building and highlighting the centrality of theoretical and empirical considerations in parameterization schemes, and in particular assumptions about scale separation, I have provided an account of structural uncertainty that starts to address epistemic and representational problems facing scientists and philosophers as well as the socio-political needs of policymakers. This is not a complete account, but it highlights how a different strategy for answering philosophical questions can provide insights into scientific practice that have not been highlighted enough so far.