1 Introduction

Recent analysis of understanding has shown that it comes in degrees (Baumberger, 2019; Khalifa, 2017). As such, understanding and its negatives form a contrary, not a contradictory category. Thus, mapping the spectrum of understanding is not a trivial task. For instance, some accounts have offered suggestions about the evaluation of degrees of understanding (Baumberger, 2019; Kelp, 2015; Khalifa, 2017; Petkov, 2021), and specified some of the conditions of understanding, approximate understanding, and the failures to achieve understanding (Park, 2017, 2020). Given this rich literature and its recent developments, we believe that we can present an updated and unified account focused on explanatory understanding, and then use such an account to present a more systematic and expanded study of the conditions under which explanatory understanding cannot arise. The conditions of such failures to achieve understanding are of primary interest to us, and their analysis is the central goal of the present study.

Here we limit our analysis to cases of understanding that depend on explanations, because both the logic and pragmatics of explanations as well as the general features of explanatory understanding are all well mapped (Khalifa, 2017). Therefore, for the most part we will base our analysis on cases that can be covered by classical accounts of explanations. This narrow scope might initially seem limited, but we believe that having a clear foundational analysis of explanatory understanding and the conditions under which it cannot arise, will offer a solid starting point for discussing more complex scenarios, such as cases of explanatory understanding derived from models that assert a level of idealization. Therefore, we focus on explanations based on classical theories of explanations in the main text. In the appendix, we shall suggest how our analysis can be extended to broad considerations about understanding involving modeling and idealization which have been hotly debated in recent literature.

To warrant the claim that an analysis of the failures to reach understanding is philosophically and methodologically interesting, we distinguish between genuine understanding,Footnote 1 and two types of epistemic failures: lack of understanding and misunderstanding. Whilst some of the existing literature has touched on lack of understanding (Khalifa, 2013, 2017; Park, 2017), no analysis of the difference between misunderstanding (as we define it) and the lack of understanding yet exists.

In Sect. 2, we roughly chart explanatory understanding, based on existing accounts, as requiring subjective and objective components. We describe its objective component in terms of a correct explanation in Sect. 2.2. This correctness is often determined by specific theories of explanations that cover structural, informational, and ontological requirements for explanations. The subjective component for us can be captured by the cognitive abilities and skills necessary to grasp or construct a correct explanation, plus the concomitant phenomenon of insight (the “aha!” moment) or a gradual increase of satisfaction (Sect. 2.3). Finally, we frame explanatory understanding as the alignment between the subjective and objective components (Sect. 2.4).

Using such a general theory of explanatory understanding, in Sect. 3 we systematize the existing claims about the lack of understanding. These concern situations where the explanation is incorrect, or situations where the cognitive abilities and skills of the explanation seekers are insufficient to grasp or construct a correct explanation.

We suggest an addition to the existing views on explanatory understanding. We hypothesize that the epistemic success of understanding gives rise to epistemic inclinations. We describe an epistemic inclination as the subjective tendency of the habitual employment of previously proven to be successful, explanatory strategies. This addition serves to highlight our characterization of misunderstanding.

In Sect. 4, we offer (what we believe) is our original contribution to the literature: An analysis of the notion of misunderstanding. This concerns cases where one’s epistemic inclinations do not align with an otherwise correct explanation. Therefore, misunderstanding is distinguished from a lack of understanding by the rejection of a correct explanation in the presence of skills and abilities necessary for its grasping.

In Sect. 5, we illustrate the cases of misunderstanding by offering a meta-philosophical analysis of the debate around the existence of distinctively mathematical explanations.

In Sect. 6, we conclude with the idea that the presence of epistemic inclinations in scientific or philosophical debates does not necessarily lead to dogmatism.

Finally in the Appendix, we offer some suggestions on how our analysis can be extended to more complex cases, such as explanations that contain idealizations, as well as understanding that seems to directly stem from modeling without the presence of an explanation.

2 Explanatory understanding

2.1 The subjective and objective components of explanatory understanding

Recent work on understanding shows that it is a broad epistemic category (Baumberger et al., 2017). Thus, an exhaustive analysis of understanding and its negatives would extend beyond the scope of a single paper, or at the cost of analytic precision. To avoid this, here we aim to obtain a good sketch of explanatory understanding and the conditions under which it cannot occur. This will enable us to use it as a starting point for developing a more comprehensive picture in the future.

Another difficulty is that, besides the numerous discussions of different sources of understanding, the concept of understanding itself is hotly debated too. In order to avoid sidetracking, we strive to provide an account that is maximally compatible with the existing views. To achieve this, our first step is to split the notion of understanding as having two components. Typically, explanatory understanding is seen as requiring the right kind of explanations and the right kind of epistemic subjects (Kelp, 2015). Therefore, its components can be presented as objective and subjective ones.Footnote 2 Using this distinction, our first sketch at a definition of explanatory understanding is:

The epistemic achievement that results from (subjectively) grasping or constructing the (objectively) correct explanation for a given fact.

To clarify, (for more details see Sect. 2.3), subjective grasping involves the abilities and skills to “assimilate” the explanatory information or construct an explanation. How this assimilation can be assessed, as for instance by the sort of inferential “tests” that Grimm (2010) or Newman (2014) posits, is something that we leave aside. Subjectively constructing an explanation for us simply means that the epistemic agent suggests an explanation to herself, or to another explanation seeker, or to a relevant scientific community.

What we call an objectively correct explanation, does not necessarily mean anything like the realist claim that the world contains explanations for its facts, and these need only to be verbalized. By “objectively correct” here we merely mean that an explanation is a type of discursive artifact—an artifact that serves a communicative purpose (Potochnik, 2017; Wright, 2012).

As such, once the explanatory seeking problem is formulated clearly (i.e., conceptualized as a specific why-question), and the explanation is presented, we are left with a linguistic or conceptual construct of some kind. In the simplest case, this structure involves a question and an answer (van Fraassen, 1980); but more often in scientific contexts, it is represented by some sort of inference (Kitcher, 1989; Petkov, 2015). Consequently, this discursive construct can be evaluated in terms of its structural and informational properties.Footnote 3 Such an evaluation can be made from the first-person, reflective perspective of the creator of the explanation, or from another—the explanation seeker. Even more broadly, an evaluation of an explanation might come from a whole epistemic group (i.e., a scientific community). All in all, the epistemic success of understanding can be seen as an alignment between these subjective and objective components.Footnote 4

To further clarify on these, we first focus on what correct explanations are in Sect. 2.2, and then move on to what grasping or constructing such explanations entails in Sect. 2.3.

2.2 Explanations and the objective components of understanding

In this section, we analyze the requirements for explanations from two sides: theories of explanations, and theories of understanding. Starting with the theories of explanations, our goal here is not to develop an original theory of explanation. Therefore we assume explanatory pluralism as a foundational claim. On this view, there is no dominant way in which scientific explanations are successful in producing understanding. Moreover, each theory of explanations has its own requirements (Khalifa et al., 2020; Pincock, 2018; Rice & Rohwer, 2021), for what counts as a correct explanation. In addition to this plurality of explanatory approaches, recent theories of understanding argue that there are different criteria for evaluating the content of explanations (de Regt, 2017, 2023; Khalifa, 2017, 2023).

Taking a hint from Khalifa (2017), we classify these requirements as:

  1. 1.

    Structural requirements, postulated by theories of explanations;

  2. 2.

    Informational requirements, postulated by theories of explanations;

  3. 3.

    Ontological requirements for the content of explanations, postulated by theories of understanding.

We characterize explanations as correct, when they meet all the criteria above; and incorrect when they fail to meet some. However, criterion 2 has a complication. It can be under- or over-satisfied. Explanations that under-satisfy criterion 2 count as incomplete, as they do not provide sufficient information for solving the explanatory seeking problem. Consequently, such explanations also count as incorrect, but this must be taken narrowly in relation only to the specific informational requirements of the explanatory seeking problem. On the other hand, explanations that over-satisfy criterion 2 count as excessive, as they provide more than sufficient information. By the same token, excessive explanations are incorrect as well.Footnote 5 The rest of this section is dedicated to clarifying these requirements.

Typically, theories of explanations provide two types of requirements for explanations. The first can be labeled as structural. It describes the form that the explanation should take, relevant to the type of explanatory relation it instantiates. The second can be labeled as informational. It encompasses the requirements for the type and the amount of information that the explanation should exhibit.

Regrettably, here we can only clarify these criteria through some rough, typical examples. We found it hard to provide a unified and completely explicit account, because each theory of explanations provides specific requirements for both the structural and informational properties of its type of explanations. To put it differently, explanatory pluralism postulates that different possibilities of explanations exist, because each theory of explanations suggests (often subtle) differences in both the structure and informativeness of explanations.Footnote 6

Nevertheless, starting with structural requirements, these can be very minimal in scale, such as those discussed by van Fraassen (1980) who characterizes explanations as direct answers to why-questions. Typically, however, explanations take the form of an inference from the explanans to the explanandum. In such cases, the structural requirements overlap with the formal requirements for such inferences plus extra requirements for the type of explanations the theory covers. For instance, in deductive-nomological explanations (Hempel, 1965), explanations take the form of a deductive inference, with a general premise that is a lawful generalization, plus relevant auxiliary conditions; and the explanandum must figure as a conclusion from them. In distinctively mathematical explanations (Lange, 2013), the structure is similar; except that the generalization is a description or a proof that some property is instantiated in a given mathematical structure. That structure then is linked with /or represented by/ some of the properties of the explanandum. Finally, the explanatory conclusion is that the property in the explaining mathematical structure holds also for the explanandum (See Sect. 5).

Causal explanations come in various types as well. They can take the form of a causal narrative (Currie & Sterelny, 2017), or a description of a mechanism (Craver, 2007; Craver & Darden, 2013), or a conditional (Lewis, 1973). In the case of counterfactual causal explanations for instance, the antecedent of the conditional is the cause (explanans), the consequent is the effect (explanandum), and the relation between the two is that of causal dependence (conceived as necessity or probability) (Lewis, 1973; Woodward, 2003).

Of course, explanations can also display complex structural properties by being both mathematical and causal (Craver & Povich, 2017). Some have suggested that explanations by unification can also be seen as cases of relating deductive and causal information (Petkov, 2015).

Despite this plurality, theories of explanations share the common ground that, if the structural requirements for an explanation are violated, the resulting explanation would be incorrect. For example, an incorrect deductive explanation could be an invalid deduction; or an incorrect causal explanation could postulate spurious causal links, etc.

These structural requirements are complemented by informational requirements, as theories of explanations also provide criteria for the informativeness of the explanans. This covers the right kind and amount of information in the explanation, relevant to the requirements of the specific explanatory seeking problem. Judging the right kind and amount of information depends on both the explanatory seeking problem and the type of explanation sought. When the informational requirement is violated, the resulting explanation can be incomplete, excessive, or simply inadequate. To paint a clearer picture, we provide an illustration inspired by van Fraassen’s pragmatics (1980) and Potochnik’s (2017) account of causal explanation patterns. Given a specific explanandum, the following three explanations can all count as incomplete, excessive, sufficiently informative or even inadequate:

  1. (A)

    All salt dissolves in water, because CH: salt is sodium chloride, and sodium has loosely bound outer electrons and is therefore highly reactive;

  2. (B)

    All salt dissolves in water, because CH, and STR: but a mass of granular soluble substances dissolves more quickly than a mass of cubes of the same substance.

  3. (C)

    I bought salt cubes, because they were on sale.

Given an explanatory seeking problem “Why does salt dissolve in water?” A would be sufficiently informative, B excessively informative, and C inadequate. On the other hand, if the question is “Why does a cube of this substance dissolve more slowly than a heap with the same mass?”, A would be incomplete, B sufficiently informative, and C inadequate. Finally relative to “Why do we have this salt that dissolves so slowly?”, A would be inadequate, B incomplete, and C sufficiently informative.

This framing of the objective components of explanations allows us to investigate explanations as clearly defined explanatory seeking problems, and adequately evaluate the informativeness of the explanatory answers. Consequently, this permits us to say that, at least in some situations once a clear explanatory seeking problem is proposed there are cases of informatively correct, informatively incomplete, or excessive explanations, relative to that problem.

Here we add that the structural and informational components of explanations are also intimately linked. This is because, very often scientific problems are open-ended. In this sense, it is possible to use more than one explanatory strategy to solve the same problem. A classic example is Salmon’s balloon case (Salmon, 1990). Where the explanatory problem “Why did the balloon fly towards the cockpit of a plane during liftoff?”, can receive a causal or a general theoretical (i.e., via unification) explanatory solution. However, in order for any of these solutions to be counted as correct, they essentially impose a structural constraint on the explanandum, by interpreting some of its aspects. This postulates a more concrete explanatory seeking problem. Consequently, specific causal or unificatory information would count towards a correct solution. Importantly each of these solutions is judged via their structural and informational properties, relevant to the theory of explanations, which covers them. We return to this issue in our discussion on understanding and inclinations (in Sect. 2.4). But for now, it is important to establish that the structural and informational requirements for an explanation taken together can lead to asserting a narrowly correct explanation, even where a plurality of other explanations can be suggested.

The often open-endedness of scientific problems also suggests that, the explanatory seeking problem can require a structurally complex explanation, that synthesizes structural and informational requirements from more than one theory of explanation. Such explanatory projects can thus impose their own constraints on what counts as structurally and informatively correct explanation. This can be illustrated again by the debate between the accounts of distinctively mathematical and causal explanations in our case study. As we shall see in Sect. 5, proponents of distinctively mathematical explanations claim that, for cases of distinctively mathematical explanations, the addition of causal information leads to excessively informative explanations; whilst proponents of causal accounts claim that distinctively mathematical explanations are incomplete, because a correct explanation should be a combination of mathematical and causal information.

To summarize, relative to the informational requirements of explanations, the sufficiency is determined in relevance to the specific explanatory seeking problem, and the theories of explanations that cover it. An explanation can be inadequately informative and thus incorrect; incomplete when the information is not sufficient; and excessive, when the information is more than sufficient.Footnote 7 Finally, an explanation can be correct when the structural criteria are met and the information in the explanation is sufficient.

Turning our focus to the ontological requirements, theories of understanding also dispute the criteria for judging the truthfulness of explanatory information. This has been at the center of the so-called factivity debate. According to factivists, the ontological status of the information in an explanation should be true in a correspondence sense. This means that the truth makers of the information in the explanans must be facts (Wilkenfeld, 2017). Non-factive accounts (de Regt, 2017; de Regt & Gijsbers, 2017), on the other hand, claim that explanations can fall short of being true in this correspondence sense. As an alternative, they claim that the explanans can be true, based on a truth maker that is something other than a straightforward correspondence with facts.

To provide more details, Wilkenfeld (2017) holds that understanding is tied to truth in terms of representational accuracy, assuming a correspondence theory of truth. For non-factive accounts such as de Regt’s (2015, 2017, 2023) pragmatic approach, the evaluation of the explanans can involve a variety of criteria, such as its intelligibility, effectiveness in promoting predictions, practical applications, or its general heuristic value (de Regt & Gijsbers, 2017). As such de Regt (2015), for instance, suggests that the evaluation of the explanans could still be true, while the truth maker is due to some other criterion, besides factual correspondence. For instance, he promotes the idea that a pragmatic theory of truth can function as such an evaluation (de Regt, 2015).

This debate has further spilled out to considerations about scientific realism vs instrumentalism, as well as the role of idealization in explanations (see the Appendix). To avoid sidetracking here we adopt the solution proposed by Khalifa (2023).

Given the plurality of judgments for evaluating the appropriate type of explanatory information, Khalifa (2023) has described the identification of truth makers as imposing further ontological requirements on explanations. In order to transcend the factivity debate, Khalifa has suggested a reconciliation between the two opposing positions, by promoting what he calls “explanatory voluntarism” (2023, p. 45). This means that theories of explanations provide the structural and informational criteria for explanations, whilst the ontological requirements on explanations coming from theories of understanding need not be committed. The ontological requirements can be a simple correspondence with facts for factivists, or a more complex set of criteria, such as the pragmatic ones promoted by non-factivists. According to Khalifa’s explanatory voluntarism, both factivists and non-factivists can agree that correct explanations provide understanding, while disagreeing on the ontological requirements for correctness. Thus, for instance both camps would agree that explanations derived from idealized models provide understanding, but disagree on how to evaluate the ontological status of such information. For factivists these explanations are true or approximately true in the correspondence sense, whilst for non-factivists their evaluation depends on pragmatic truth or other contextual criteria (see the Appendix).

Without delving further into the factivity debate, we adopt Khalifa’s explanatory voluntarism. This enables us to claim that when the ontological requirements for an explanation are not met, the explanation fails to be correct. In addition, we propose that both structural and ontological requirements are also critically involved in providing understanding via correct explanations. We believe that this move sets the bar for understanding from a theory-neutral perspective. It in turn advances our discussion of lack of understanding as resulting from incorrect explanations.

2.3 The subjective components of explanatory understanding

Here, we explicate the subjective dimension of understanding as the epistemic achievement (1) supplemented by a specific phenomenal hue, and (2) requiring cognitive abilities and skills. This setup has a theoretical motivation, as the literature mostly discusses the subjective side of understanding under the concept of grasping. Grasping in turn has two sides—phenomenological and epistemic.

Psychologically, grasping has been discussed as being accompanied by a specific phenomenal hue—the so-called “aha!” feeling or an insight moment (Baumberger et al., 2017). The literature converges on the idea that the subjective feeling of understanding is neither necessary nor sufficient for understanding (de Regt, 2009; Hills, 2009; Lipton, 2009; Trout, 2002; Ylikoski, 2009), due to the unreliable nature of the feeling itself. This is certainly the case for the typical analysis of understanding. But we do not think it shows that the phenomenological dimension of understanding is of no theoretical importance, especially when we consider explanations and the relevant epistemic outcomes such as understanding, lack of understanding, and misunderstanding. For instance (as we shall discuss in Sect. 4), the absence of the feeling of understanding can lead to rejecting an otherwise correct explanation, even when the cognitive abilities related to reaching understanding via such an explanation are present. This motivates us to include the phenomenological side of understanding into our analysis.

Understandably, most philosophical analyses of understanding in terms of grasping emphasize the epistemic side, whilst underestimating the phenomenological side. This is perhaps because the initial efforts were focused on clarifying what kind of epistemic success understanding is. But we believe that theories of understanding are now mature enough to be extended to provide a more detailed analysis of the phenomenological side of understanding. Also, it seems to us that at least some specific epistemic failures depend on such a phenomenology. Therefore, here we extract the claims on the epistemology of grasping and combine them with the phenomenology of understanding.

The epistemic side of grasping can be broadly characterized as an assimilation of the explanatory information into one’s belief system. This assimilation has been described in numerous ways. For instance, Khalifa (2017) characterizes grasping as the agent’s capacity to assimilate an explanation within her belief system which potentially contains other relevant explanations of the phenomenon. Others (Grimm, 2010; Newman, 2014) analyze grasping as the ability to use the explanatory information to construct inferences similar to the one exemplified by the explanation. Grasping can be embedded into theories of explanations as well. For instance, under the manipulationist accounts of causation, grasping is the ability to “see” how a manipulation of the explanans would lead to a change in the explanandum, etc. (Woodward, 2003).

All these accounts share the common ground that grasping requires a number of cognitive abilities and skills. However, they diverge on what such abilities and skills are, and if they provide a distinct epistemic achievement, different from knowledge. In other words, they try to answer the question: “Is grasping the correct explanation simply equivalent to knowing the correct explanation?”.

This problem is at the center of the debate between knowledge-based accounts—which classify understanding as a type of knowledge (Grimm, 2006; Kelp, 2015; Khalifa, 2017; Mizrahi, 2012; Strevens, 2017), and skill-based accounts—that distinguish understanding as a distinct epistemic category (de Regt, 2017, 2023; Newman, 2014), characterized by the development of skills. In order to clarify the role of cognitive abilities and skills for grasping, we will shortly overview the debate between de Regt and Khalifa.

For the skill-based accounts championed by de Regt (2017, 2023), understanding is a distinct epistemic category, because it requires the development of context specific skills. These skills can be minimally the skills required to make the explanation intelligible to the explanation seeker. An explanation is intelligible to a subject if she can solve similar problems as the one presented by the explanation, or more broadly if she “sees” how the explanatory relation works. That is to say, for instance one’s deductive reasoning skills are required to understand the explanandum by a deductive-nomological explanation, because they are at the heart of solving the explanatory seeking problem. For de Regt, such skills are separated from what is required for believing and thus simply knowing the correct explanation.

Opposing this view, the knowledge-based accounts downplay the importance of such skills. These accounts do not deny that some abilities and skills are involved in grasping. However Khalifa, for example, argues that these are of a rather trivial kind. This is because scientific knowledge already “requires a fair amount of ability” (Khalifa, 2017, p. 54). Moreover, according to him, these abilities are reducible to knowledge-that, and thus do not exceed knowing (Khalifa, 2012). Therefore, concerning the same case of deciding if a student understands the explanandum through a given explanation or is merely able to recite the explanation, Khalifa makes the same judgment. However, his assertion that the student lacks understanding is based on the fact that the student does not possess the knowledge of deduction, that concerns the inferential connection between the explanans and the explanandum (Khalifa, 2012, p. 26).

To clarify, for both de Regt and Khalifa, to grasp an explanation one also needs to be clear on how the inferential relation between the explanandum and the explanans works. Both Khalifa and de Regt essentially make the same judgment in the case of the student who merely “knows by heart” the correct explanation—that she does not understand. The essential difference, however, is whether she lacks knowledge or skill. That is to say, for Khalifa this student lacks a crucial piece of knowledge about the inferential connection between the explanans and the explanandum. For de Regt on the other hand, grasping this inferential information is an acquired cognitive skill, separate from the information of the explanans itself.

We need not settle the dispute here, as for now it has been largely resolved by an academic interchange between de Regt and Khalifa in a recently published work (Lawler et al., 2023). There, de Regt acknowledges Khalifa’s friendly effort to incorporate the essential use of skills in establishing the inferential relation between the explanans and the explanandum. Thus, it is safe to say that for both accounts the use of cognitive abilities and skills plays a central role for archiving the epistemic success of grasping. The flip side is that the lack of cognitive abilities and skills would thus lead to a lack of understanding. We follow this common ground and incorporate the lack of cognitive abilities and skills as one source of lack of understanding, illustrated in more detail in Sect. 3. Here we highlight that the lack of abilities and skills is not the only contributing factor for the negative dimension of understanding. As we shall see below, there are other factors (one of which we characterize as epistemic inclination in Sect. 2.4) that can lead to what we shall call misunderstanding, illustrated in more detail in Sect. 4.

To take stock, in this section we characterized the subjective components of understanding by combining the epistemic and psychological sides of grasping. In so doing, we defined subjectively grasping an explanation as:

(A) requiring abilities and skills to assess and evaluate how the explanation offers a solution to the explanatory seeking problem; and (B) having a positive phenomenal reinforcement of an insight of how the explanation solves the explanatory seeking problem.

Our next step is to combine the subjective and objective components of understanding, and provide a more explicit definition of explanatory understanding in Sect. 2.4. Based on that definition, we present the negative dimensions of understanding—lack of understanding in Sect. 3, and misunderstanding in Sect. 4.

2.4 Explanatory understanding and epistemic inclinations

In order to link the subjective and objective components of understanding, without having to develop a complete theory of explanations, an exhaustive account of grasping, and the role of abilities and skills, we found it useful to think about explanation as a type of problem-solving cognitive task.Footnote 8 This can be the analytic task of assessing and evaluating the information in an existing explanation, or arriving at a novel one.

In the former case, an explanation is a type of input. That input can be seen in some cases as requiring an analytic solution. For instance, when the cognitive agent is evaluating an explanation with a familiar structure and sufficient explicitness. In this case, grasping the explanation would be equivalent to analytically assessing and accepting it as correct. In other cases, this input can require a specific insight; for instance, when the explanation involves an unfamiliar, novel mode of reasoning, or is not sufficiently explicit. In this case, grasping the explanation would be equivalent to filling the informational gap within the explanation, or mastering the novel mode of reasoning exemplified by that explanation.

An explanation can also be a type of output. This occurs in cases where no previous explanation is available, and a novel one has to be developed. This again can be an analytic task, where the epistemic agent’s previous knowledge is sufficient to solve a familiar problem. For instance, applying an explanatory pattern serially to cover phenomena within its scope (Grimm, 2010; Kitcher, 1989; Petkov, 2015). An explanation can also require a completely novel insight, in cases where a novel explanation must be discovered.

To generalize over these tasks, we can categorize them as requiring a gradual analytic solution, an insight, or both. As such, the objective component of the quality of the explanation, the employment of subjective cognitive abilities and skills, and the phenomenal hue of understanding would all be relevant for the resulting epistemic achievement. Given such framing, we find that relevant research from cognitive science and psychology bears on accurately linking the subjective and objective components of understanding.Footnote 9 As we shall see, this will also offer a previously unexplored dimension of understanding, as a process that forms epistemic inclinations.

To begin, both the literature on problem solving (Kaplan & Simon, 1990; Kounios & Beeman, 2014; Osuna-Mascaró & Auersperg, 2021) and on understanding (Baumberger et al., 2017) emphasize on the presence of some kind of phenomenology. This can be a form of insight, or a gradual increase of satisfaction from an analysis. Insight is seen as a sudden change in the type of knowledge representation or in the formation of a new concept relevant to the solution of a problem (Kounios & Beeman, 2014; Osuna-Mascaró & Auersperg, 2021). This sudden change is phenomenally accompanied by the “aha!” excitement (Kaplan & Simon, 1990; Kounios & Beeman, 2014; Osuna-Mascaró & Auersperg, 2021). Analytic solutions in turn are seen as the solver’s deliberate manipulation of the “problem elements to discover the solution” (van Steenburgh et al., 2012, p. 475). This in turn is accompanied by a gradual, incremental warm feeling of satisfaction (Kounios & Beeman, 2014).

Of course, much the same way implicit learning can occur without phenomenal awareness, it is possible that understanding can happen without the supplemented meta-cognitive feeling (Cleeremans et al., 1998; Trout, 2002). However, we believe that explanatory understanding might be an exception to this rule. This is because explanations are conceptually explicit solutions to cognitive problems (i.e., explanatory seeking ones). Such solutions can be discovered (requiring an insight “aha!” moment) or analyzed (requiring deliberate analysis and a gradually increasing satisfaction of arriving at the correct solution). This makes arriving at explanatory understanding a more clear-cut epistemic effort. Consequently, the conceptual or discursive artifact of an explanation has to be in some sense explicitly present and complete. This artifact, as we noted (see Sect. 2.2), has to fulfill a given number of structural, informational, and ontological requirements, in order to solve the explanatory seeking problem. For such a completeness to be asserted, the epistemic agent has to have some kind of “break point” mechanism which would indicate that the construct satisfies the required structural, informational, and ontological properties imposed by the explanatory seeking problem. This “break point” signifies that the epistemic agent does not need to engage with the explanatory seeking problem anymore.

Following the research of Ylikoski (2009) and Keil (2003), we believe that the phenomenal dimension of insight of understanding plays exactly this role. According to them, the sense of insight plays an important cognitive role, because it facilitates cognitive economy. Since we are limited epistemic agents, the pleasurable feeling of satisfaction that one reaches from a sense of insight, indicates that an epistemic goal has been achieved. Thus, such a feeling acts as a breakpoint, by signifying that the subject has received/or reached/sufficient information and no longer needs to cognitively engage with the problem. Similarly, an analytic solution can vary in its explicitness and depth, and consequently also requires some kind of cognitive mechanism that would indicate that a solution has been reached.

This role of a breaking mechanism (we argue) is played by a phenomenal sense of insight, or gradual satisfaction from a solution. Moreover, the relation between the positive phenomenology and epistemic economy, implies another important finding in cognitive science of problem solving—insight is more easily reached when one solves problems that call for familiar solutions (Lovett & Anderson, 1996; Schultz & Searleman, 2002). This preference for familiarity manifests itself also as a tendency to apply familiar solutions to novel problems (Lovett & Anderson, 1996; Schultz & Searleman, 2002). This in turn could lead to positive or negative epistemic outcomes. Crucially and in similarity to research in epistemology, a sense of insight can be misleading.Footnote 10 For instance, recent studies of Ammalainen and Moroshkina (2021) and Grimmer et al. (2022) have shown that triggering a false insight is possible by semantic priming or misleading hints. Some (Grimmer et al., 2022; Salvi et al., 2015, 2020) have interpreted these results as indicating that insight phenomenology functions to draw cognitive resources (attention) to ideas that are most consistent with one’s implicit knowledge. Therefore, if past knowledge contains false beliefs, the resulting insight could be incorrect as well (Laukkonen et al., 2020; Webb et al., 2019). This interpretation of insight is also consistent with the paradigm which interprets insight as an implicit process of reasoning (Bowden, 1997; Grant & Spivey, 2003; Grimmer et al., 2022; Hattori et al., 2013; Laukkonen et al., 2018; Salvi & Bowden, 2020; Salvi et al., 2015; Schunn & Dunbar, 1996; Sio & Ormerod, 2009).

What this suggests to us is that, the sense of insight functions as an indicator of conformity between relevant, previous knowledge and novel knowledge. More narrowly for explanatory understanding, this suggests that the sense of insight can also function similarly. It will bias the epistemic agent to assert an explanation that conforms more easily with previously accepted explanations. The function of the sense of insight then is, to indicate that the explanation can be asserted, based on its conformity with the structural, informational, and ontological properties, that other already familiar explanations display. Obviously, such previously acquired explanations might vary in quality and thus the sense of insight would undoubtedly be judged epistemically, as a fallible indicator of understanding (Grimm, 2009). Again the emphasis is that, the sense of insight on its own is not to be blamed as being misleading, but that the implicit information that triggers it has epistemic flaws.Footnote 11

However, for us it is more important to establish that, there is a tendency in epistemic agents to apply or receive insights from familiar solutions. And thus, they are more inclined to use such solutions to solve novel problems. This tendency to seek or assert explanations that display structural, informational, or ontological properties that are familiar or embodied by previously successful explanations, we call epistemic inclinations.

To develop this idea further, we establish that on the subjective side explanatory understanding requires:

  1. (1)

    the relevant cognitive abilities and skills necessary to analyze or reach an insight to the correct explanation;

  2. (2)

    this cognitive process is supplemented by a specific phenomenal hue.

    To this we add that:

  3. (3)

    the epistemic success of understanding also brings about the tendency of the habitual employment of previously proven to be successful, explanatory strategies.

We further hypothesize that the cognitive and psychological basis of such inclinations lie in the notion of mental sets which is closely related to problem solving (Schultz & Searleman, 2002). Mental sets are the tendencies of cognitive agents to apply previously successful solutions to problems that seem familiar (Lovett & Anderson, 1996). Mental sets are typically related to impairments of successful problem-solving tasks (Schultz & Searleman, 2002). This is because they lead to ignoring possible but unfamiliar solutions, in favor of familiar ones. A good illustration is an experiment of chess problem-solving, carried out by Bilalić et al. (2008). In the experiment, chess players were required to find a checkmate position with the fewest number of moves. Players were given a problem that has two possible solutions: one with five moves and the other only three moves. If the players were familiarized with the five-move solution, they would tend to use it, due to familiarity, instead of opting for the optimal but unfamiliar three-move solution.

Of course, solving scientific problems does not appear as clean-cut as checkmate scenarios which present definitive problems and solutions. As we mentioned, in scientific contexts explanatory seeking problems can be open-ended. However (in Sect. 2.2), we suggested that once certain structural, informational, and ontological requirements are imposed on a scientific problem, that problem can receive a relatively narrow set of correct solutions. We suggest that mental sets play a part in the formation of such narrower interpretations of scientific problems. The reason is that they involve a specific way of looking at the explanatory problem itself and trying to solve it using previously familiar strategies.

To further unpack the notion of epistemic inclination, we define it as: (1) a way to frame the explanatory problem based on familiarity, which leads to (2) asserting explanations (by analysis or discovery) that conform to specific structural, informational, and ontological requirements.

To give a simple illustration, if one has an inclination to seek causal understanding, when confronted with an explanatory problem, she would tend to (1) frame the problem as a causal one, and (2) try to provide a causal explanation that conforms to her implicitly/or explicitly/accepted theory of causality. This in turn would also postulate certain structural, informational, and ontological constraints on the evaluation and acceptance of potential explanatory solutions.

We do not claim that epistemic inclinations are completely equivalent to mental sets, much less to mental rigidity. In our case, we describe mental inclinations as the tendency of applying previously successful explanatory strategies to novel problems, and as such potentially facilitating or hindering understanding. This idea is also in tune with the notion of grasping, which captures the epistemic success of understanding as the ability to employ the explanation to solve similar problems (see Sect. 2.3 above). To add we will suggest that epistemic inclinations are subject to change, due to deliberate rational belief revision (in Sect. 6).

For now, it is important to emphasize that such inclinations can lead to a positive or negative epistemic outcome. It is negative in the sense that an agent might reject alternative explanatory strategies, exactly because they do not trigger a positive phenomenology. A phenomenology in turn depends on one’s implicit epistemic inclinations. Consequently perhaps, the notion of epistemic inclination merits its own specific study, but here our goal is to provide a bedrock analysis of both understanding as forming such inclinations and their role in reaching understanding, lack of understanding, or misunderstanding.

With this we can expand our basic definition of understanding to:

The cognitive achievement that occurs from the subjective grasping or construction of a correct explanation, that requires specific cognitive skills and abilities, is often accompanied by positive phenomenology, and leads to the development of relevant epistemic inclinations.

With this definition in mind, we can examine the epistemic phenomenon of understanding as being a consequence of an alignmentFootnote 12—an alignment that occurs between the subjective and objective components of understanding. This alignment entails (1) the right abilities and skills of an epistemic agent, plus the positive reinforcement role of the phenomenal sense of satisfaction or insight, and (2) the correctness of the explanation itself.

In the context of explanatory understanding, the alignment of its components leads to the agent solving the explanatory problem to her satisfaction and thus to the epistemic success of understanding. Conversely, misalignment leads to the problem remaining unsolved. This can occur in two different situations. Firstly, when any of the components is not in order, such as on the subjective side when the agent lacks abilities and skills, or on the objective side when she grasps an incorrect explanation. Or secondly when there is a misalignment between the components themselves, in cases where a correct explanation engages an agent who in principle has sufficient abilities, skills, and knowledge to grasp an explanation, but rejects it because it does not conform to her epistemic inclinations. The former case, as we shall see shortly, leads to a lack of understanding, whilst the latter to misunderstanding.

3 Lack of understanding

To shift our attention to the negative dimension of understanding, we first characterize cases of lack of understanding, as situations where people fail to solve the explanatory seeking problem they have set. For us, all cases of lack of understanding are cases where no correct explanation is constructed or grasped. This implies that a lack of understanding can occur when:

  1. (1)

    subjectively the agent lacksFootnote 13 the cognitive abilities and skills necessary to grasp or construct a correct explanation; or

  2. (2)

    the agent grasps an objectively incorrect explanation—an explanation that fails some of the three requirements outlined in Sect. 2.2.

On the subjective side, when one’s cognitive abilities and skills are inadequate to grasp or construct a correct explanation, one lacks understanding. These cases are widely discussed in the literature, typically by the skill-based accounts. In Sect. 2.3, we illustrated this type of epistemic failure by the example of a student who lacks the reasoning skills and simply memorizes the explanations in the textbook. Relative to the phenomenology of understanding, the positive feeling of understanding does not occur in cases that fall under this category.

It is also possible that the agent grasps an incorrect explanation. Such an explanation can fail any of the three requirements for correct explanations: structural, informational, or ontological.

Starting with explanations that are structurally incorrect, as we mentioned such explanations can fail some of the structural requirements of the relevant theory of explanations. For instance, an agent can grasp a mere valid deductive explanation that fails to exemplify a lawful generalization; or a causal conditional that is based on spurious causes.

When the informational requirements are violated, we encounter a more complicated scenario. This is because the criterion which determines if an explanation is informatively adequate is contextually sensitive. It means that the satisfaction of the informational requirements of an explanation depends on the specific requirements of the explanatory seeking problem itself. We tried to provide such a contextually sensitive criterion by defining explanations as informationally sufficient, inadequate, incomplete, or excessive. Importantly, as we showed in Sect. 2.2. the same explanation can turn either way by simply switching or redefining the requirements of the explanatory seeking question. Having in mind such contextual sensitivity, we can nevertheless generalize about the understanding that such explanations will provide. Obviously sufficiently informative explanations generate a solution to an explanatory seeing problem and as such count as correct, and thus providing understanding. The negative cases however are more complicated.

Whenever an explanation is informatively inadequate, it wholly misses the mark of the informational requirements of the explanatory seeking problem, and as such does not generate understanding. A myriad case of transitivity of causality in the literature of causal explanations can serve as a good example. Reiss (2015), Northcott (2008), and Hitchcock (2003), for instance discuss the following narrative: “Captain yells ‘fire’, trainee fires. Upon hearing the command, victim ducks. The bullet misses the victim who survives unscathed.” For the question: “What saved the victim’s life?”, one answer could be: “The victim survived, because the captain yelled fire.” The literature on causal explanations converges on the idea that such an explanation misses the mark of informativeness and as such leads to a lack of understanding.Footnote 14

An even more complicated scenario occurs when explanations are incompletely or excessively informative. Here we defend the claim that an incomplete explanation provides some explanatory information, but it does not lead to an increase of understanding. This is because the explanation fails to meet the exact requirements of the explanatory seeking problem. However such an explanation, if taken regarding a different but related explanatory seeking problem, can be seen as satisfactory, and therefore as increasing one’s understanding. For instance, if the explanatory seeking problem a scientist has set is to understand the exact mechanism of long Covid complications, she can fail the task with an incomplete explanation that “Covid19 causes long Covid”. This explanation is incomplete, because the scientist is at least clear that Covid19 is the main cause of long Covid. Hence, the offered explanation does not increase her understanding. For, she has failed to solve the problem of describing the mechanism of long Covid. However, for the question: “What causes such a condition?” the answer “Covid19 causes long Covid” would be sufficiently informative and increase her understanding.Footnote 15

Finally, excessively informative explanations present an even more curious case. On the one hand, it is possible that the explanation seeker can manage to simply “cut off” the informational content at the right place, and as such reduce the excessive explanation to a sufficiently informative one. Again provided with a detailed explanation of the mechanism of long Covid, one can manage to extract information for the reduced explanation “Covid19 causes long Covid”, for her problem “What causes long Covid?”. On the other hand, the added complexity of such explanations might make it difficult to understand the explanation itself, and as such a failure to understand the explanation would also lead to a lack of understanding. Using the same example, one might not manage to extract the explanatory information needed to solve the problem.

Moving to the ontological requirements for an explanation, the situation seems to be, at least on the surface level, similar to that of structural requirements. If the ontological requirements for an explanation are not satisfied, that explanation consequently would not generate understanding. However, in scientific contexts the informational and ontological requirements can pull in different directions. For instance, if one’s ontological requirements are strictly factive but the informational requirements are less stringent, one is often left with a choice between asserting a generalization that is strictly speaking factively false for the explanandum, and seeking to provide an overly complicated explanation. The debates between quasi-factivists and factivists, along with the studies of the role of idealization and modeling in science, provide abundant illustrations. However, we must note that the debates about both the factivity of explanations and the role of idealization in modeling take a significant portion of the contemporary literature in philosophy of science, and therefore we cannot provide a serious, systematic study here.

However, our intuition is that in many scientific scenarios, the complexity of jointly satisfying both the informational and ontological requirements for an explanation can present a challenge. Therefore, some trade-offs are often made. We shall return to these issues later (see the Appendix), but for now we can provide a simple example. The informational requirement for solving a causally complex problem might be such that, if one loosens her ontological requirements, the problem might be solvable. For instance, Potochnik (2017) has suggested that this is the case when one asserts, what she calls, causal-pattern explanations. Such explanations describe causal relations, but also assume a level of idealization. Nevertheless, Potochnik notes that these explanations generate understanding.

On the other hand, if one asserts a rigid factive ontological requirement, it will weigh on the information of a potential causal explanation to such an extent that a complete factive causal explanation might not be achievable for human agents.

With these considerations in mind, our goal here is to explore a hitherto unrecognized possibility—situations where the epistemic agent’s skills or abilities are sufficient for the task of grasping what can be asserted as a correct explanation, but she rejects that explanation nevertheless. We postulate that this rejection can occur, because the agent’s epistemic inclinations are incompatible with such an explanation. In such a case the agent might seek an alternative that fits better her epistemic inclinations. We categorize such and related cases as situations of misunderstanding.

4 Misunderstanding

Misunderstanding has two common usages in the English language: failures to understand correctly, and disagreements between two parties. We find these lexicographical definitions useful, and our use of the word “misunderstanding” in this paper is populated by them. However, we must clarify the first usage (“failures to understand correctly”). As noted above, we are inclined to delineate between “lack of understanding” and “misunderstanding”, at the point where a lack of understanding concerns epistemic failures that result in or from an incorrect explanation. Misunderstanding, on the other hand, is the rejection of a correct explanation, in the presence of the cognitive abilities and skills to grasp it. In this sense, misunderstanding can be seen as a negative propositional attitude the agent has towards a correct explanation. Consequently, the epistemic outcome from such a propositional attitude can be seen narrowly as negative (i.e., as a lack of understanding), because when one misunderstands the explanation, she does not understand the explanandum through that explanation.Footnote 16

This rejection can occur whenever there is a misalignment between some of the structural, informational, or ontological features of the explanation and the inclinations of the epistemic agent for them. As we mentioned (see Sect. 2.4), epistemic inclinations are tendencies to assert or apply familiar explanatory strategies. These strategies might have been successful in solving previous explanatory seeking problems, but that does not necessarily mean that they will always be applicable to novel ones. Consequently, and in similarity with the suggestions from the literature on problem solving, explanatory problems which require unfamiliar solutions would be typically harder to solve. We suggest that this difficulty might be due to the fact that the epistemic agent’s inclinations are ill-suited to address the specific explanatory problem.

A further negative outcome might occur in cases where there is a misalignment between the features of an otherwise correct explanation, that the epistemic agent considers, and her inclinations. Consequently, even though the agent has the cognitive abilities and skills to evaluate the explanation, as a correct solution of a specific explanatory seeking problem, the positive outcome of asserting the explanation might be blocked by a lack of positive insight. We hypothesize that this lack of positive phenomenology plays an important contributive role for misunderstanding. This is because the phenomenology of understanding or insight and the formation of epistemic attitudes towards an explanation are intimately linked. As we mentioned, the literature on problem solving and insight seem to suggest that the phenomenology of insight is a result of an implicit cognitive process of evaluation of a solution to the problem, in light of relevant previous experiences. This evaluation is suggested to be based on conformity (Grimmer et al., 2022). Concerning the narrow case of explanatory seeking problems, this suggests that the implicit evaluation of the explanation is based on such conformity with familiar types of explanations and their structural, informational, or ontological properties. This evaluation is often implicit, as the literature suggests (Grimmer et al., 2022; Laukkonen et al., 2021). Consequently, only the outcome, as the presence of a positive feeling of an insight from the given solution (an explanatory one in our case) and the tendency for its acceptance, is available to the epistemic subjects. This leads us to believe that the feeling of insight can be seen as reinforcing one’s attitude towards an explanation, even when we are engaged with an explicit evaluation of the features of the explanation. Having these ideas in mind, we can provide a definition of misunderstanding as:

Misunderstanding is the epistemic failure, that occurs when a correct explanation is offered, that explanation can be in principle grasped (as the epistemic abilities and skills of the explanation seeker are sufficient for its grasping), but understanding of the explanandum through that explanation is blocked by a mismatch between the features of the explanation and the epistemic inclinations of the explanation seeker.

To illustrate this definition, we suggest a few rough examples of misunderstanding. These examples show a mismatch between the properties of the explanation and the inclinations of the explanation seeker.

For instance, Jansson (2014) argues that the debate between Leibniz and Newton, is partially a debate about the correct theory of causation and if explanations of gravitational phenomena, as provided by Newton, are instances of causal explanations. According to Jansson, this historical debate has a modern twist, because it is investigated differently by two contestants for what would count as the correct theory of causal explanations—those of Woodward (2003) and Strevens (2008). Arguably then both theories explicate and develop different perspectives about what causality is, and form a different inclination about the structural, informational, or ontological properties of causal explanations. We suggest that, this line of dispute is possible exactly because each side addresses causal problems, by interpreting them as structural explanations within their respective theory, and thus providing informational and ontological requirements for what would count as a correct explanation. This results in rejecting their opponent’s explanations as structurally incorrect, insufficiently informative, or ontologically lacking. We propose that such a genuine scientific and philosophical dispute would be solvable, if both sides could find convergence on the structural, informational, or ontological properties of causal explanations.

We suggest that a similar scenario occurs within the dispute between proponents of distinctively mathematical explanations and causal explanations. An inclination for a deductive solution to an explanatory problem might lead to the rejection of a causal solution, or vice versa. This is because the two sides differ in their inclinations towards specific explanatory structuresFootnote 17 and their informativeness. As we shall see in the case study (Sect. 5), the proponents of each kind of explanations claim that their rivals have a misunderstanding (as we define it) of what the correct explanation would be. To use our terminology, this misunderstanding concerns a case where proponents of distinctively mathematical explanations claim that their rivals misunderstand and assert an excessively informative explanation. Proponents of causal explanations, similarly, claim that their rivals misunderstand and assert an insufficiently informative explanation.

Finally, people might diverge on their requirements for the ontological properties of an explanation as well. Some might identify the truth makers with factual correspondence, whilst others might embrace factivly weaker, pragmatic considerations. This in turn, we suggest, can lead to a disagreement over the role and level of idealizations that are acceptable in explanations, derived from idealized models (see the Appendix).

In all these cases, however it is important to note that the rejection of what can potentially be a correct explanation is not due to a lack of abilities or skills to grasp it. In all these cases, the epistemic agents involved are trained and respected scientists or philosophers for whom it can hardly be claimed that they lack specific abilities or skills to evaluate the explanations in question. Instead, we claim that epistemic inclinations, the intuitions they create, and the subsequent biases in favor of specific explanatory strategies all play a major role in such disagreements.

The picture we have painted so far might seem to lead to a perspective-dependent and perhaps relativistic account of understanding. However, very often participants in such disputes share a common scientific goal and consequently, convergences are found on some structural, informational, or ontological requirements for asserting one type of explanatory solutions over others. As such the logical outcomes of such disputes can be of two kinds:

  1. (1)

    Pluralistic where: the debaters can agree that both explanatory strategies provide correct explanations but relative to different structural, informational, and/or ontological framings of the explanatory seeking problem. During the disputes, both parties were misunderstanding each other’s otherwise correct explanations.

  2. (2)

    Monistic where: the debaters can agree that one of the strategies has been misguided, and the other is the singly correct explanatory strategy. During the disputes, the party in error was misunderstanding their rival’s correct explanation.

To both outcomes the role of epistemic inclinations is similar to that of mental sets. They would facilitate understanding through familiar explanations and hinder epistemic success in the face of unfamiliar ones. We hypothesize that both the role of mental sets in hindering or facilitating insights, and the need for cognitive economy are relevant for such outcomes. Similar to the literature on insights and mental sets, we postulate that this could be both a positive and a negative phenomenon. Negative in a sense that it could lead to a disagreement between scientists on two otherwise genuine explanatory strategies, or to make one inclined to reject an otherwise correct explanation. But positive in the sense that it could lead to the discovery of complementing explanations, or even to a broad synthesis between what were seen as initially incompatible explanatory strategies (see the Appendix).

5 Causal VS distinctively mathematical explanations: a case study

The debate on distinctly mathematical explanations concerns whether explanatory understanding of some ontic facts can be achieved through purely mathematical framing of these facts, without implicit or explicit information on the causal genesis of the represented facts. Proponents of distinctively mathematical explanations claim that, in special cases where the representative mathematical structures constrain the phenomena, information about such mathematical structures can provide understanding for the represented facts (Lange, 2013, 2017). Proponents of causal or broadly ontic accounts of explanations reject this possibility (Bangu, 2021; Craver & Povich, 2017). Instead, they claim that an explanation would be in important ways incomplete without the presence of relevant causal or broadly ontic information.

Albeit ongoing, this debate in philosophy of science is of significant interest to us. This is because the two sides argue about essentially the same factual explananda, but disagree on which type of explanations provides understanding for these explananda. Moreover, it can hardly be claimed that any side lacks the skills or abilities relevant to grasp the opponent’s arguments and explanations. Therefore, we propose that the debate can be conceptualized as an interesting case of misunderstanding where epistemic inclinations might be partially responsible for the ongoing disagreement.

Our central claim is that the two parties in the debate have different epistemic inclinations (to seek mathematical or causal understanding respectively). As we postulated in the previous section, such a disagreement may have two potential outcomes: pluralistic or monistic. In case of a pluralistic outcome, the two parties would agree that both their explanatory strategies are correct. However, during the debate when the parties were actively disagreeing with each other, they also misunderstood the other’s otherwise correct explanations. If the latter monistic outcome occurs, a single correct explanation would be agreed upon. In this case one of the sides would concede that its explanatory strategy was inadequate for solving the problem. Thus, during the debate the party in error was misunderstanding their rival’s correct explanation.

Before presenting the debate itself, we make some clarifications on our motivation and position. Firstly, our aim here is not to present a knockdown argument for any side. Instead, we analyze the debate as it stands—an ongoing debate within philosophy of science. Our motivation to choose this debate, as an illustration of our framework, is exactly due to its present state of being undecided. This essentially permits us to use it as an illustration of the two possible outcomes we described above.Footnote 18 With this we intend our analysis to serve as an illustration of a general methodology—an illustration of how our account of misunderstanding can be applied to clarify positions in tenacious disagreements that can conclude with any of these logical possibilities.

Analyzing the debate from the perspective of pluralism might at first seem like a more natural solution. However, we shall see that a deeper discussion on the ontological requirements for both kinds of explanations can lead us to assert that, the monistic reading is a viable option as well.

To begin, the canonical case around which the discussions evolve is: A parent who has twenty-three strawberries and three children cannot distribute the strawberries evenly without cutting any (strawberries!) (Lange, 2013).

Here the proponents of distinctively mathematical explanations claim that a correct explanation would be: “Because 23 is non-divisible evenly by three and this mathematical fact constrains the possibility of dividing the strawberries evenly” (such as Lange, 2013). While the proponents of causal explanations usually propose that a correct explanation would have to be: “Because the physical system is causally constrained in such a way that both the facts of having 23 strawberries and 3 children hold, the fact that 23 is non-divisible evenly by three explains the failure of the parent.” (such as Bangu, 2021).Footnote 19

If one frames the problem in two ways, it is easy to see how both parties can be right:

“Why does the parent fail to divide 23 by 3 (strawberries) evenly?” and “Why does the parent fail to physically divide 23 by 3 strawberries evenly?”. These two different framings can be further analyzed as a case of diverging epistemic inclinations. As defined in Sect. 2.4, an epistemic inclination is: (1) a way to frame an explanatory problem based on familiarity, and (2) the tendency to assert explanations (by analysis or discovery) that conform to familiar structural, informational, and ontological requirements.

The first one shows that a mathematically inclined explanation seeker, perhaps due to her familiarity and training in solving mathematical or purely theoretical problems, would tend to frame the facts of the explanandum in a formal context. The second posit entails that subjects would also seek an answer that is conceivable within a narrow theoretic context. In such a context an explanation, similar to a mathematical proof, would be preferable. They would thus tend to assert an explanation that takes the form of a deduction and involves a mathematical proof. A relevant and telling quote from Lange’s book which seems to suggest this idea is:

“By the same token, “word problems” in mathematics textbooks are full of allusions to facts that have distinctively mathematical explanations…” (2017, p. 25).

On the other side of the debate, proponents of causal accounts might be inclined to address the same problem with a different cognitive habit. Undoubtedly causal reasoning is our most common method of coping with the surrounding environment. For instance, Woodward (2003) claims that some animals and human infants develop a concept of causality, because such a concept is important to predict what would happen if they interacted with the environment in a specific way. Similarly, ontic accounts of explanations usually claim that the central aim of scientific endeavors is to predict and control the changes in our environment. This ensures the dominant role of causal explanations.

Framing the debate in this way shows that both parties have different epistemic inclinations. These inclinations underlie the way they formulate different explanatory seeking questions relative to the same explanandum. Following this line of reasoning also allows us to view each side as claiming that the other provides excessively or incompletely informative explanations.

From the perspective of mathematical explanations, adding further causal information leads to excessively informative explanations. Lange (2017) for instance seems to defend this line of argument. He claims that the addition of causal informationFootnote 20 to a distinctively mathematical explanation is explanatorily inert and does not contribute to the understanding of the explanandum. On the other hand, a causal reading of the explanatory seeking question dictates that purely mathematical answers are incompletely informative. For instance, Bangu (2021) and Bueno and Colyvan (2011) claim that whilst mathematical truths have a contributive role to an explanation, an additional causal fact plays the actual explanatory role (e.g., the law of conservation of mass).

One can be tempted to read the debate under a pluralistic framework, especially given the fact that perhaps each party’s intuitions are informed by equally important cognitive abilities (that to deduce and generalize; or that to orient oneself in a world of causes and effects). However, a deeper examination of the debate shows that the challenges from both sides are also about incompatible ontological commitments. That is to say, each party’s inclinations also concern different requirements for the ontological properties of explanations. If one pursues this interpretation of the debate, perhaps one would reach a monistic conclusion.

For instance, Lange (2017, 2021) claims that explanations in physics are organized in a nested hierarchy of necessities where higher order mathematical facts have a priority by constraining ontic facts. This then, according to Lange, makes one committed to deeper claims about the ontological status of mathematics. Essentially his argument is that both nominalism and Platonism fail to account for distinctively mathematical explanations. Instead, such explanations are possible only if Aristotelian realism is true (Lange, 2021). According to it, mathematics concerns mathematical properties possessed by physical systems, and as such mathematical properties can constrain physical ones (Lange, 2021). This claim obviously touches on the ontological status of mathematical entities and is made even more pressing, because the argument of Lange (2017) is aimed at explaining the role of essential explanatory tools of physics such as the Lorentz transformation.

Such an ontological commitment to the status of mathematical facts cannot be jointly true with the opposite claim coming from causal accounts. For instance, Bueno and Colyvan (2011) defend an indexing account of the role of mathematics in science. According to it, mathematical objects are not existing entities. This mismatch is further aggravated by the fact that both ontologies are not compatible.Footnote 21

Given the mismatch between the two parties’ inclinations about the ontological status of explanations, the solution can only be a monistic one.Footnote 22 One ontology about the status of mathematical objects and their role in science has to prevail. Consequently, the party in error has to consent that they have offered an ontologically incorrect explanation. They should also acknowledge that they have misunderstood their rivals correct explanations.

To conclude, here we would not develop further arguments that might tip the scale for any side of the debate, or for a pluralistic reading. Instead, this discussion should serve as an illustration of how our account can be applied to similar disputes about scientific explanations. As we showed, it could be used to analyze the outcomes where one side prevailed, or if they ended up in mutual recognition.

6 Conclusion. epistemic inclinations impasse or progress?

It might seem that we have painted a rather grim picture for solving tenacious debates in philosophy and science. However, one needs to recall that we developed the notion of epistemic inclination in relation to mental sets, but different than mental rigidity.

For us, epistemic inclinations result from receiving a positive phenomenology that occurs after a successful problem solving. As we mentioned, the cognitive science of insight suggests, that insight phenomenology functions to “select” ideas from the stream of consciousness by drawing attention to ideas that are most consistent with the agent’s implicit knowledge (Salvi et al., 2015, 2020). This makes the available background knowledge essential for the development of inclinations. Importantly however, our background knowledge is often subject to rational belief revision. It is possible that the introduction of new arguments and ideas can result in a revision of one’s implicit belief system. Therefore, as in any debate with the introduction of further arguments and sufficient open-mindedness from both parties, a consensus might be reached. This is because each party’s inclinations about the structural, informational, and ontological features of what constitutes a correct explanation would become explicit. Given such explicitness, one can also revise or further affirm their epistemic inclinations in light of new information. Consequently, a consensus can be reached. Such a consensus would occur if the parties reach an agreement on the structural, informational, and ontological properties of the potential explanatory solutions to the scientific problem they seek to resolve. Therefore, we believe that scientific misunderstanding due to diverging inclinations could also lead to progress.