Potentiality realism: a realistic and indeterministic physics based on propensities

We propose an interpretation of physics named potentiality realism. This view, which can be applied to classical as well as to quantum physics, regards potentialities (i.e. intrinsic, objective propensities for individual events to obtain) as elements of reality, thereby complementing the actual properties taken by physical variables. This allows one to naturally reconcile realism and fundamental indeterminism in any theoretical framework. We discuss our specific interpretation of propensities, that require them to depart from being probabilities at the formal level, though allowing for statistics and the law of large numbers. This view helps reconcile classical and quantum physics by showing that most of the conceptual problems that are customarily taken to be unique issues of the latter -- such as the measurement problem -- are actually in common to all indeterministic physical theories.


I. INTRODUCTION
Centuries of formalization of physics have led to a narrative according to which the universe supposedly evolves through deterministic laws, namely, every event happens as a necessity.This is considered almost a truism in classical physics (Newtonian mechanics, electromagnetism, relativity) and even after the advent of quantum mechanics, many consider the probabilistic prediction of the theory as a mere epistemic concept, while at the fundamental level everything is governed by the deterministic Schrödinger equation.Perhaps due to such a strong belief of most of the physicists in determinism, it seems that indeterminism has assumed the status of a bug to be eliminated, and this has led to a series of deep-rooted misconceptions that tend to scramble indeterminism together with a lack of causality or realism, or of lawfulness altogether.
We defend that it is among the essential characteristic of science to provide explanations of natural phenomena, namely the possibility of telling a story about how Nature does it.And this requires to introduce metaphysical elements that causally interact, which provides a strong motivation for (some form of) realism (see Section V).Moreover, we are quantum physicists, and our positions are strongly influenced by the indeterministic worldview brought about by quantum theory, a view rooted in fundamental results such as Heisenberg relations or the violation of Bell inequalities. 1We have also gone beyond quantum theory, noticing that believing in determinism even at the classical level is too strong of an assumption, which is not supported by observation and requires to assume infinite information at every point of space(-time).In fact, in a series of recent works, we have proposed alternative, fundamentally indeterministic interpretations of (classical and relativistic) physics [1][2][3][4][5][6].
The aim of this paper is therefore to spell out in some detail what a realistic description of an indeterministic world en-tails.We will show that there is a natural way of maintaining both realism and indeterminism if one assumes that physical systems are characterized at the fundamental level not only by their actual possessed properties, but also by their intrinsic (quantifiable) potentialities. 2Our view drastically differs from traditional realistic positions which uphold objects to be existing and identified by their intrinsic actual properties (at any given time).We maintain instead that physical objects exist and are still characterized by their intrinsic properties which are, however, in general only potential (at any given time), i.e. intrinsic tendencies to actualize.Our view can therefore be called potentiality realism. 3

II. PROPENSITIES IN AN INDETERMINISTIC WORLD
A main feature of an indeterministic world is to be able to produce non-necessary events, hence to produce new information, i.e., information that was not existing and that is truly created as time passes. 4This new information could be merely random, devoid of any structure [12].Indeterminate events, however, have in general well-determined potentialities of various strengths to realize one out of all the possible results, when they eventually become determinate (examples of possible results are spin up or spin down of a spin-1/2 quantum particle, different faces of a dice, bits 0 or 1 of a random number generator, etc.).Hence, such a description requires objective, causally determined -and yet not ruled by necessity-intrinsic tendencies for individual events to obtain, i.e., propensities. 5We deem propensities to be as necessary in a description of a causal indeterministic world as actual possessed properties are necessary to describe a deterministic world.
In the quite extensive literature devoted to propensities, there have been several proposals to conceive these objective, intrinsic tendencies.In what follows, we will expound our interpretation of propensities, arguing that this is the most suitable one to describe indeterminism in physics while maintaining realism and causality.
To begin with, our view departs from a major category of propensity theories pursued by several authors, such as I. Hacking [17] or D. Gillies [15], who advocate that propensities are "long-run", i.e., they are dispositional properties of long, possibly infinite, series.If this were the case, however, it would entail a sort of "nonlocality in time", because if a propensity is to be an element of reality, then it would have to exist as a single entity that influences series of events happening over a span of time that could require the entire age of the universe and its (possibility infinitely extended in time) future.Hence, we reject this view of a time-nonlocal infinite sequence of runs which determines the behavior of individual events, but rather the long-run observed frequencies are manifestation of individual-event propensities.Longrun sequences are nothing but the accumulation of individual events, while the opposite does not follow.In fact, propensities should characterize objective tendencies of events that can in principle happen only once and not be repeatable (in which case however they are not even approximately measurable).Note that in the case of a single observed outcome, the only information that can be extrapolated about the associated propensity is that it was different from zero.
On the contrary, we define a propensity to be an intrinsic property of objects -which in physical theories are called systems-to realize a particular (idealized) measurement outcome, i.e., a quantified tendency to realize a particular outcome when subjected to specific circumstances.The specific circumstances are the complete set of causally relevant conditions.Our interpretation of propensities is therefore "single-case", and represents a form of indeterministic causation aimed at quantifying a degree of objective possibility (which makes propensities almost an interpretation of probability but with due conceptual and formal distinctions; see Section III).On this account, our view is similar to the early ones of J. H. Fetzer [19] and of L. Ballentine [18], the latter of whom maintains that a "propensity refers to a degree of causality that is weaker than determinism."[18].Propensities are therefore to be thought of as "indeterministic laws" or "potential forces", really existing in the world and causally connecting events but in a looser sense than necessity (which would be the case in a deterministic world).
More formally, a propensity is described by a mathematical entity (called itself a propensity) that -given two causally related events (e.g., the hypothetical outcome of a measurement with its relevant preparation)-quantifies the strength of the tendency for the effect to happen, given its causally relevant conditions.In this sense, even if the world is characterized by indeterminacy at the fundamental level, events are related by causality.Indeterminism does not come out of the blue, is not acausal but is the consequence of fundamental indeterminacy. 6n Section III, however, we will see that retaining this causal interpretation of propensities would require them to slightly modify their formalism with respect to mathematical probabilities.And yet, while they may not be probabilities at the formal level, they share most of the mathematical structure of probabilities, therefore one can still adopt the writing: p(a|causal conditions).This is to be interpreted as the intrinsic tendency of the event a to happen (for instance, the outcome of a possible measurement performed on a system) given the relevant causal conditions that can influence the realization of that outcome (for instance, the settings of an idealized measurement). 7But while a is an event in mathematical terms, i.e., an element of a probability space, the causal conditions are not necessarily so.This is clear in quantum mechanics, where the causal conditions to compute the outcome a of a measurement (i.e., an eigenvalue of the observable A, corresponding to the eigenstate |a ) are represented by the quantum state ψ -which is a ray in a Hilbert space and not an element of the probability space-and the choice of measurement A; the probabilities are then computed according to the Born rule as p(a|ψ, A) = | a|ψ | 2 .These "probabilities" arising in quantum mechanics are, in fact, sometimes called "generalized probabilities" (see e.g.[25], chapter 8.1) because they are not defined on standard sets of events, as in Kolmogov's axiomatization, due to the incompatibility of canonically conjugate variables (i.e. the algebraic structure of the set of "quantum events" is non-Boolean).Hence, "the functions assigning probabilities to quantum events are, paradoxically, not probability functions at all, at least, not in a Kolmogorov's sense."[25].In the next section, we will discuss in detail the difference between propensities and probabilities and their analogies with quantum "generalized probabilities".
Furthermore, contrarily to the standard view of (at least classical) physics, the outcomes are not a manifestation of actual pre-existing properties that get unveiled.In that view, classical states (as well as realistically interpreted quantum states) are collections of actual values (mathematically associated to n-tuples of real numbers) that are assumed to ex-ist well-determined at any instant in time, no matter how far in the past or in the future.In our view, on the other hand, outcomes are dynamically realized -i.e., they come into existence, thus creating new information-from an array of mutually exclusive potential outcomes.Obviously, welldeterminate outcomes do exist (and every measurement yields one and only one outcome), but only as a particular case when the value of the associated propensity is 1, i.e., when no alternative is possible and the outcome is deterministic (see Section III).As such, at any instant in time before the event that corresponds to the realization of one (and only one) outcome out of all the possible ones, the element of reality -i.e. the state-is the collection of propensities associated to each possible and in general still indeterminate outcome.Our proposal is therefore a form of realism that can be called potentiality realism as opposed to the standard form of actuality realism.The element of reality is now conceived as a collection of potentialities, each of them quantified by an objectively existing propensity. 8Let us emphasisze that propensities are possessed by the system in the same way as actual properties are.
Our definition of states in terms of propensities requires some clarification.J. Butterfield defines a physical state "as a system's maximal (or 'complete') set of intrinsic (or 'possessed') properties" [20].Our definition of a state complies with this.Complete means that a state encapsulates the maximal amount of information existing at present about each relevant degree of freedom of the system.But the maximal information is here about potential properties, i.e. the collection of all the propensities.Note that thinking in terms of propensities does not add to the "ontological cost" of a theory (see also Section V): propensities are indeed not directly epistemically accessible, but neither are the postulated actual values of physical properties in standard views of physics.In both cases, the knowledge is obtained through the same operational procedure of repeated measurements and collected statistic that supposedly approximate the underlying element of reality, be it an actual value or a potential one.
Let us turn now to the second desideratum of a physical state, namely that of being a set of intrinsic properties.Intrinsic means in this context that the propensities are properties of the system alone and are not characteristics of experimental arrangements [13], nor of chance set-ups [17].D. H. Mellor states that attributing propensities to a chance set-up "is to remove completely the point of ascribing a disposition as something that is present whether or not it is being displayed.It is to confound propensity with the chance distribution that displays it" [21].In fact, intrinsic tendencies (i.e., possessed propensities) can only be approximately revealed by ideal measurements, in the same way as possessed properties are (like the value of a physical quantity, e.g., the energy of an atom).Ideal measurements are noise-free measurements.In the standard view of classical physics, they are characterized by yielding the same result without any variance (because they merely reveal the actual value).Now, let us assume that it is always 8 A similar position was previously elaborated in Ref. [27].
possible to find a state that makes the propensity for a certain measurement outcome on a system to take the value 1 (i.e., they yield the result deterministically) and that an ideal measurement of a propensity is independent of the value of the propensity itself.Then one can use the state corresponding to a deterministic outcome (propensity 1) to characterize the ideal measurement also for any other state.
Note that, additionally, the collection of all the propensities -i.e., a state in this "potentiality" interpretation-may have its own well-defined mathematical structure, hence it might be possible to summarize this collection of propensities in a compact mathematical form.
Finally, it ought to be recalled that the violation of Bell's inequalities -by now a well-corroborated empirical fact-forces us to reject locality.This obviously affects propensities as well, which have to be in general nonlocal elements of reality.Namely, a single objective tendency quantifies the bias towards the possible realization of correlations between distant indeterministic outcomes (a more detailed discussion can be found in Ref [4], also in relation to special relativity).

III. PROPENSITIES VS. PROBABILITIES: HUMPHREYS' PARADOX
In the previous section, we have defined propensities as being intrinsic, objective tendencies that causally and indeterministically quantify the possibility of a system to realize an event, given its causally relevant conditions.This seems prima facie exactly what a conditional probability describes, if one is to interpret uncertainty not as epistemic but as a manifestation of fundamental indeterminacy.However, a result known as "Humphreys' paradox" [26] has posed serious limitations to directly interpreting propensities as objective probabilities.Ballentine rightly calls "Humphreys's result a theorem, rather than a paradox because it is a result that validly follows from its assumptions."[18].In fact, we contend that this can be casted in the form of a no-go theorem which states that the conjunction of the following assumptions is untenable (i.e., leads to contradiction): (i) Propensities quantify indeterministic causal connections.
(ii) Propensities are probabilities, i.e., they are defined through all Kolmogorov's axioms (in particular those from which Bayes' rule is derived).
To see this, assume that propensities are indeed probabilities -assumption (ii).Consider two causally related events, namely, if one of them -the cause, C-obtains, it influences the tendency of another event -the effect, E-to obtain.This means, by assumption (i), that there is a propensity P(E|C) that quantifies the tendency of E to happen given that C has occurred.Let this propensity connect the event to the cause non-trivially, i.e., P(E|C) = P(E).At the same time, since we have assumed propensities to encapsulate the concept of causal connections, whether the effect E will happen or otherwise (i.e., the complementary event E happens), it cannot influence the cause, i.e., P(C|E) = P(C|E) = P(C). 9Note that this is the standard definition of causality -also knows as no-signaling from the future-in operational formulations of physical theories (such as in "generalized probability theories", GPTs [31]), too, where the choice of a measurement cannot influence the state preparation (see Ref. [32]).By assumption (ii), Bayes' rule ought to hold.This means that one can write the conditional probability in terms of its reversed one as (Bayes' rule): However, plugging in the causality constraint above yields to P(E|C) = P(E), which is in direct contradiction with the assumption of non-triviality of the cause.
Dropping assumption (i) would lead to accept that propensities are not single-case tendencies but long-run (as defended by Hacking [17] and Gillies [15]), but we have already argued against this stance in the previous section.On the other hand, if one wants to maintain assumption (i), propensities need to escape Humphreys' theorem by departing from formal probabilities.Axiomatizations of propensity calculus -not reducible to Kolmogorov's axioms of probability theory (or equivalent)-have been proposed by Fetzer and Nute [19,28], one of the present authros (NG) [29,30], and Ballentine [18].
Note that if propensities are to be compatible with statistical observations, their axiomatization cannot be fully arbitrary and consequently they cannot be arbitrarily different from probabilities, for, in turn, the latter are also constructed to be the limiting case of observed frequencies.That is why Gillies proposes "to speak of the probabilistic causal calculus [i.e., axiomatized propensity calculus] as a non-Kolmogorovian probability theory by analogy with non-Euclidean geometry."[15].In fact, in the same fashion that non-Euclidean geometry maintains most of the structure of standard geometry by only rejecting Euclid's fifth postulate, propensity calculus (non-Kolmogorovian probability) drops the Kolmogorov's axiom(s) that lead to the derivation of the Bayes' rule.In particular, as already remarked in Ref. [18], a desideratum for propensities (which is also true for probabilities) is that they obey Bernoulli's Law of Large Numbers: for every ε > 0 and for the limit of n → ∞.In words, P is the probability that quantifies the correlation between the relative frequency of the occurrence of the considered outcome, f n , and the propensity p for that event, which gets stronger as the number of trials n increases. 10This result is derived in standard probability theory, but there p is taken to be a probability as well instead of a propensity.Bernoulli's law of large numbers is therefore at least a necessary element in common between probabilities and propensities, for both of them should be the limit of relative frequencies in long series of measurements.However, at the conceptual level the difference is tremendous: while probabilities only encapsulate correlations (which do not imply causation), propensities cause relative frequencies, which, in turn, are the observable manifestation of propensities and therefore the operational way to epistemically access them (to an arbitrary approximation).This is why Bayes' rule is valid for probabilities but not for propensities.
Another desideratum for propensities -also in common with probabilities-is that they ought to be bounded between 0 and 1, because they have to account for impossibility and certainty, respectively.But again, due to their causal nature, a propensity of 0 or 1 does not only mean certainty -of failure or occurrence of the considered event, respectively-in a measure theoretic sense (as it is in probability theory, where events with probability 0 are not logically excluded but their occurrences form a subset of measure zero).Rather, a propensity of 1 means a necessary causal connection, i.e. deterministic causation between the conditions and the occurrence of an event.On the other hand, an intermediate value of a propensity, strictly larger than 0 and smaller than 1, should be instead interpreted as a non-deterministic "potential force".A propensity with value 1/k represents a truly unbiased random event for the property of a system that can display k mutually exclusive outcomes.
Since, however, measurements always lead to a single observed outcome, this necessarily begs the question of how the potentialities become actual, i.e. what "mechanism" makes propensities evolve from an intermediate value, between 0 and 1, to either 0 or 1 at the time of measurements (and not necessarily only at measurements).This is exactly the analogous of the notorious "quantum measurement problem" (in Refs.[1] and [5] we discussed this calling it the "classical measurement problem"), which is a general characteristic of all fundamentally indeterministic theories.At the formal level, in quantum mechanics the measurement problem can be cast as an incompatibility between the linear, unitary, deterministic evolution of the quantum state (Schrödinger equation) and the observation of a single outcome upon measurement with the "collapse" of the quantum state on one of the eigenstates of the operator corresponding to the measurement.Note that in an indeterministic interpretation of classical physics this can also be regarded in the same terms: the indeterminate state (i.e., the collection of propensities) is bounded by a finite region of phase space which gets deterministically mapped, through the equations of motion (Newtons laws), to an arbitrarily large region of phase space (due to chaotic systems).When a measurement is performed, however, a localization is expected in the same fashion of a "collapse" in quantum physics.Hence, in both (indeterministic) classical mechanics and quantum mechanics this actualization of the potentialities requires to look for a "mechanism" that "forces" the propensities to get determined at measurements.The two possible mechanisms that can lead to this is that propensities either evolve into determined outcomes spontaneously -which is reminiscent of objective collapse models in quantum theory [33,34]-or if a measurement somehow "imposes" a determination -similarly to the Copenhagen interpretation of quantum mechanics in which an observer or a measurement apparatus (not described within quantum theory) makes the wave function "collapse".The former view seems the most compatible with reductionism, whereas the latter one seems to imply a form of a top-down causation [35], where a higher level of description "forces" the actualization of the lower level.Furthermore, the fact that only one outcome can be realized in a given experiment is itself a (metaphysical) assumption.In fact, in principle one can think of a radical interpretation of propensities, in which the possibilities never actualize and the state is the collection of the all the propensities for any possible event of the universe and extended everywhere in time.This is the analogous of the "many-worlds interpretation" of the quantum state and it shares with it its enormous ontological baggage (see Section V).Finally, we notice that the standard deterministic interpretation of classical physics is analogous to adding unobservable "hidden variables" (i.e., the real numbers as the actual value of physical variables, see [3]), in the same way that Bohmian mechanics adds hidden variables to the quantum state (i.e. the position of each quantum particle) to deterministically complete the theory.
Upon repetitions, under the same causally relevant conditions, one can measure with arbitrary precision the propensity of a system, and approach it in the limit.If the value of the propensity is 1, the system will always display the same particular outcome -under the assumption of ideal measurements (see Section II).Note that -according to standard classical physics-to measure the supposedly existing actual value of some physical quantity of a system or to measure the (also supposedly existing) propensity for a system to display some particular result -as postulated in potentiality realism-one has to accumulate enough statistics.Indeed, when we assert that a system possesses (at present) a certain property, e.g., that some physical quantity has some value a, we mean that if we would test for this property, i.e. we perform a (ideal) measurement of that physical quantity, we would find with certainty this property, i.e., the value a.In practice, however, such tests need to be reproduced many times until noise and false positive can be dealt with.This is very similar to the case when we assert that a system possesses (at present) a certain propensity.
It is in this statistical analysis of measurements that standard Kolmogorov probabilities play their role.This is very well understood by the mathematical measure theory (in terms of probability spaces, Borel sets, etc).However, and this is crucial, there is a priori no reason to believe that sets of statistics corresponding to different measurements can be combined into a single probability space.This is not just a theoretical hypothesis, but it is actually demonstrated by quantum theory.Hence, propensities of a system to display results corresponding to different physical quantities, i.e., different measurement settings, should not be expected a priori to satisfy Kolmogorov's axioms -in compliance with Humphreys' theorem.In fact, if there were a global probability space for all propensities p(a|measurement conditions), then the joint probability p(a, measurement conditions) would be welldefined and this would imply the existence of "hidden variables" -the elements of the global probability space-that provide a deterministic description.Then one is left with two possibilities: 1.If the hidden variables are accessible, then this world is not really indeterministic, or 2. If the hidden variables are fundamentally inaccessible, then this world would require to be described through variables that are intrinsically non-physical (like, e.g., real number in classical physics or exact positions in Bohmian mechanics).
A clarification is here in order.In the second statement, fundamentally inaccessible has a stronger meaning than the inaccessibility of actual values of standard classical physics or of the propensities in our proposal of potentiality realism.The latter two, in fact, while effectively inaccessible, can be approached by successive approximations in repeated measurements (by means of Bernoulli's law of large numbers).The hidden variables of the global probability space instead would be as inaccessible as actual positions in Bohmian mechanics, and have therefore a much less justifiable physical ground. 11

IV. FINITE INFORMATION QUANTITIES
In Ref. [1], we have proposed an alternative, indeterministic interpretation of classical physics, based on propensities.Starting by noticing that physical variables are usually assumed to have an actual value, determined with infinite precision (encapsulated by a real number), we assumed instead as a fundamental principle the finiteness of information density, namely that finite volumes of space(-time) can contain only a finite amount of information (see also [2]).We have thus introduced a model of classical physics in which physical variables are assumed to take values, instead of in the real numbers, in a new class of mathematical entities that we named finite-information quantities (FIQs).To illustrate this, let us start from the standard view.Let Γ be a physical quantity (say the position of a particle) that take values in the unit real interval, i.e., Γ ∈ [0, 1] and let us write it in binary base: 11 This treatment in terms of hidden variable has some resemblance with the solution to the Humphreys' theorem proposed by D. Miller [37] and the late Popper [38].They maintain that propensities can only be attributed to the whole causal past (the past light cone in relativistic terms or the entire state of the universe) of an event: "The non-standard conditionalization is on an 'event' outside the probability space-the entire present state of the universe-which determines the probabilities of the events in the probability space" [14].We notice that Bell nonlocality jeopardizes the view that all the relevant causal connection lie only in the past light cone of a local event.Instead, one should take into account the compound of the past light ones of all the systems entangled with the one under consideration.
where the bits γ j ∈ {0, 1}, ∀ j ∈ N + .Because Γ is a real number, its infinite bits are all given at once with a well defined value 0 or 1.In our model, on the other hand, we impose that the information should be finite, such that not all the digits should be determined at all times.However, since we require it to be empirically equivalent to standard classical physics, the first (more significant and perhaps known) n bits should be fully determined at time t: where each bit γ j ∈ {0, 1}, ∀ j ≤ n(t), and the symbol ?k here means that the kth digit is a not yet actualized binary digit, but only its propensity exists before time t.For each digit j of a physical quantity Γ(n(t)), we associate the propensity q j ∈ [0, 1] ∩ Q that quantifies the tendency of the j-th binary digit to take the value 1, such that it is q j = 1 iff the j-th bit is certainly 1 and q j = 0 iff the j-th bit is certainly not 1 (i.e. if it is 0).A FIQ is then defined as an ordered list of propensities {q 1 , q 2 , • • • , q j , • • • }, that satisfies the (necessary) condition: ∑ j I j < ∞, where I j = 1 − H(q j ) is the information content of the propensity, and H is the binary entropy function of its argument.This ensures that the information content of FIQs is bounded from above.A physical quantity Γ thus reads: Since most of classical systems are chaotic, this fundamental indeterminacy (here modeled with FIQs), inevitably leads to indeterminism in classical physics too (see Ref. [2]).Therefore, FIQs render classical physics indeterministic while upholding a concept of physical state in potentiality realism (see Section II).Indeed, a pure state (i.e. containing the maximal information possible about a system) is here a list of propensities with the constraint that the information therein contained is always finite.

V. WHY REALISM? BENEFITS AND COSTS OF AN ONTOLOGY
In the previous sections, we have defended the position that propensities -at least in their causal, single-case interpretation advocated for here-allow to maintain realism and at the same time fundamental indeterminism.One may, however, ask why one should look for a realistic account of physics in the first place. 12It should be recalled that it is generally understood that science, and physics in particular, is a form of knowledge characterized by both explanatory power and predictive power.While the latter is perhaps considered the most signature feature of science, that alone leads to a mere instrumental use of scientific theories which does not seems to bring particular insights on nature, and even precludes the meaningfulness of certain questions (a prime example is that of Bell's inequalities that arguably would have not been derived within a fully instrumental use of quantum theory).
The explanatory power is the ability to tell consistent stories about how Nature does it.It requires to charge our theories with metaphysical elements because in order to give explanations one has to postulate the existence of entities in the world that interact and causally account for the observed phenomena.Obviously, this alone would not be science, for also mythology or religion aim to explain natural phenomena, but in a more arbitrary way (for example, a stormy sea could be explained by the god Poseidon, or the thunder bolt by the god Zeus who are in anger).So, in a satisfactory scientific theory, the metaphysical elements should be built as an interplay between the observations and the predictions of a theory, on the one hand, and the underlying things of the world on the other, i.e., postulated elements of reality (these can be fields, particles, specific degrees of freedom with their actual magnitudes, propensities, etc.). 13A necessary condition for these underlying elements of reality is that, while being in general not directly observable, they are compatible with -i.e. they do not contradict-the predictions of a theory and actual observations.These things of the world form the ontology of a theory, which provides the foundations of realism at least as a working hypothesis.
Note that there are many (in principle infinite) different possible ontologies compatible with the same theory and set of observations at any given time.One should therefore define standards or guidelines to adopt one ontology among all the possible compatible ones.As examples of possible ontologies consider, for instance, the one theorized by the pre-Socratic philosopher Anaxagoras who maintained that everything is composed of "seeds" (spermata), namely microscopic and fully formed versions of any observed substance and beings (e.g., miniaturized versions of humans, stones, horses, trees, etc.) and that "in everything there is a portion of everything" [36].More recently, the so-called many-worlds interpretation of quantum theory postulated the existence of a multiplicity of (possibly uncountably infinitely many) worlds, namely that for each and every quantum measurement the whole universe, with its tremendous complexity, splits into many copies thereof that only differ from each other by the result of the quantum measurement.What unites these two ontologies is that, although they provide an explanation, they are extremely "costly", for they both imply an absurd inflation of required things of the world: Anaxagoras in terms of types of com- 13 Physicists may be acquainted with the definition of an "element of reality" as defined by Einstain, Podolski and Rosen in their influential EPR paper, in terms of perfect predictability of physical quantities.Our use of the term "element of reality" here does not refer to that definition and it simply intuitively refers to the "things" that exist in the world.
ponent, many-wolds in terms of amount of universes (so an inconceivable amount of information, etc.) that form reality.This suggests that one standard to prefer an ontology over its competitors is to favor more "economical" ontologies, given the same compatibility with the observations and predictions of a theory. 14Therefore, if the observed phenomena can be explained resorting to less types or a smaller amount of elements of reality, that ontology should be preferred. 15ow, it is generally known that quantum physics is the most successful theory and that its predictions are only probabilistic.A more thoughtful analysis of classical physics also leads to conclude that at the observational level also classical outcomes are only known with an interval of confidence that is characterized by statistical repetitions from which one extrapolates a probability distribution.And chaotic systems make the indeterminacy grow exponentially as time passes.(Note that the standard story of classical physics explains away probabilities by assuming that there exists actual, infinitely precise determinate values that are approximately revealed by measurements, but this is already an ontological stance).Therefore, in the actual practice of science, measures of indeterminacy, i.e. probabilities, play not a but the major role.Thus the most economical ontology is to assume that probabilities are manifestations of an underlying ontology that resembles them in structure, i.e. indeterministic forces in the form of causal propensities as expounded in this paper.

VI. DISCUSSION
We have discussed an interpretation of physics that we called potentiality realism.While not being tied to any particular physical theory, it helps clarify the rather generally misunderstood fact that realism and fundamental indeterminism are compatible metaphysical properties.Potentiality realism grants to intrinsic tendencies or propensities, the role of "elements of reality".A physical state is thus not characterized by its actual properties (the postulated values taken by each variable corresponding to each relevant degree of freedom), but by potential, non-deterministic "forces" that quantify the intrinsic tendency of a property (as opposed to its possible alternative) to actualize.
We would like to emphasize the picture of indeterminism in physics that our view presents.Here, indeterminism is not represented by stochastic fluctuations in the dynamical evolution, nor by arbitrarily acausal "jumps"; here no causal chain starts from nowhere.In our view, an indeterministic evolution is the causal consequence of fundamental indeterminacy.As time passes, new information is created and thus the indeterminacy is reduced.Depending on the dynamical system, the evolution is then driven by the fresh information, one way or the other.This view of indeterminism is not acausal, simply because, as time flows, some potentialities get excluded.
While our view allows to regard the world as indeterministic while maintaining a clear causal structure, a number of issues remain unsettled.For instance, how do potentialities become actual?Is it a spontaneous process, or does this pave the way to think in terms of non-reductionism in which different domains of reality need to act on each other to provide an explanation?Furthermore, in this view, states are not collections of actual values, but of propensities (i.e., potential values, of which the actual ones are a subset, corresponding to a unit propensity).But propensities dynamically change their values, eventually becoming 0 or 1 when the outcome of an experiment obtains.But then, according to what laws do the propensities evolve?Do they require a second-order dynamics?
By posing these challenging problems, potentiality realism provides a general conceptual framework to analyze the problems of quantum theory in relation to its classical counterpart.In fact, it partly deprives quantum physics of its uniqueness with respect to most of its conceptual issues and "mysteries".It shows that what is peculiarly attributed to quantum mechanics was in the large part the result of a historical contingency: Fundamental indeterminacy, the measurement problem, a plurality of interpretations (including a "many-world one" and Bohmian mechanics) are all present in a potentiality realistic interpretation of any physical theory, being it classical or quantum (or any hypothetical post-quantum theory).What remains unparalleled in quantum physics is the existence of incompatible variables, which is in fact the only place where the Plank constant h appears (and in Bell nonlocality, which requires measurements in incompatible bases to violate the classical local bound [39]).
Interpreting quantum mechanics remains one of the greatest challenges of modern science.But if one thinks twice, this challenge lies to a large extent above and beyond quantum physics and, hidden behind historically rooted dogmatisms, the great challenge has always been to interpret physics tout court.