1 Introduction

Most of the literature on belief merging concerns sources of the information of equal reliability (Chopra et al. 2006; Everaere et al. 2010a; Konieczny and Pino Pérez 2011; Lin and Mendelzon 1999). Such a scenario occurs, but not especially often. Two identical temperature sensors produce readings that are equally likely to be close to the actual value, but a difference in made, age, or position changes their reliability. Two experts hardly have the very same knowledge, experience and ability. The reliability of two databases on a certain area may depend on factors that are unknown when merging them.

Merging under equal and unequal reliability are two scenarios, but a third exists: absent reliability. Most previous work in belief merging is about the first (Chopra et al. 2006; Everaere et al. 2010a, 2020; Haret et al. 2020; Konieczny and Pino Pérez 2011, 2002a; Lin and Mendelzon 1999); some is about the second (Cholvy 1998; Konieczny et al. 2004; Lin 1996; Revesz 1997); this one is about the third.

The difference between equal and absent reliability is clear when its implications on some examples are shown.

Example

Equal reliability

Absent reliability

Two experts

They have the very same knowledge, experience and ability

They differ in knowledge, experience and ability, but how much of these they possess is unknown

Two sensors

They are of the same kind and are in the same condition (temperature sensors located next to each other, distance sensors with the same orientation)

They are of different kind, or are in different conditions, and which one is more reliable in the current situation is not known

Two databases

They cover the very same domain, and are equally likely to be correct

They cover different domains, so that a certain piece of information may have been crucial to one but a detail in the second

The assumption of equal reliability is quite strong in the example of the two experts; rather, there may be some reason to believe one more than the other; not knowing who, this scenario falls in the case of absent reliability. For the two sensors and the two databases equal reliability is not unlikely, but so is absent reliability.

If reliability is absent, can it be assumed equal?

When merging preferences, yes. When merging beliefs, no.

Merging preferences (List 2013; Lang 2004; Mata Díaz and Pino Pérez 2017) aims at obtaining a result that best reflects the collective opinion of a group. A common premise is that all members of the group have the same weight on the final decision, as formalized by the condition of anonymity. In lack of information telling otherwise, equal weights are a valid assumption.

A technical example shows why not when merging beliefs instead. Three scenarios are possible: A, B and C; two sources of information rank their unlikeliness on a scale from 0 to 3, with 0 being the most likely and 3 the least (unlikeliness scales are common in belief revision (Darwiche and Pearl 1997; Katsuno and Mendelzon 1991; Rott 2006), in spite of likeliness being more intuitive). The first source grades A as the most unlikely scenario, the second as the most likely; numerically, unlikeliness are 3 and 0. Both sources grade B as kind of likely (1), and C in the opposite way of A (0 and 3).

Scenario

Unlikeliness according

Unlikeliness according

 

to the first source

to the second source

A

3

0

B

1

1

C

0

3

Two different cases are considered: in the first case, the first source is twice as reliable as the second; in the second, no reliability information is present.

If the first source is twice as reliable as the first, the overall unlikeliness of the three scenarios are \(2 \cdot 3+0 = 6\), \(2 \cdot 1+1 = 3\) and \(2 \cdot 0+3 = 3\). The minimum is 3: the least unlikely scenarios are B and C.

Scenario

Unlikeliness according

Unlikeliness according

 

to the first source, weighted

to the second source, weighed

A

\(2 \cdot 3\)

0

B

\(2 \cdot 1\)

1

C

\(2 \cdot 0\)

3

If some fact x holds in B but not in C, its truth is uncertain since it differs in the two most likely scenarios.

The second considered case is when reliability information is absent. The table of overall unlikeliness can no longer be computed since it requires not only the unlikeliness of the scenarios according to the sources but also the reliability of the sources. A tempting solution is: “since the relative reliability of the sources is absent, it is assumed equal”. This allows to compute the table again, this time with multipliers 1 and 1.

Scenario

Unlikeliness according

Unlikeliness according

 

to the first source, weighted

to the second source, weighed

A

3

0

B

1

1

C

0

3

The most likely scenario is now B, and B only: its overall unlikeliness 2 beats those of A and of C, both 3. If the fact x holds in B but not in C, it is deemed true.

The two cases differ both in their initial information and in their conclusions. In the first case, reliability information is present, and x is not concluded. In the second case, reliability information is absent, but x is concluded. Starting from more information leads to less information.

Knowing that x is true is more information than not knowing the value of x. That the first source is twice as reliable as the second is more information than an unspecified relative reliability.

Information is not just different. It is strictly more in one case than in the other: knowing that the first source is twice more reliable than the second is strictly more information than no given reliability; knowing that x is true is strictly more information than not knowing it. The case starting from strictly more information ends up with strictly less information.

This example is inspired by the “penny z” of Popper (1959, pp. 425–426): a coin is initially assumed fair in lack of information indicating otherwise; adding the confirmation of fairness does not change its probability of falling heads or tails. In the interpretation of probability as degree of belief (Hájek 2012), the probabilities are the epistemic state. Adding information should alter the epistemic state, but the addition of fairness (which is new information) changes nothing.

This example was used by Popper against the subjective interpretation of probabilities, but relies on the principle of indifference: events of unknown probability are assumed equally probable (Keynes 1921; Shackel 2007). The Bertrand paradox (Bertrand 1889; Keynes 1921; Shackel 2007) shows it is problematic; the coin example by Popper (1959) shows another contradictory aspect of it (Gärdenfors and Sahlin 1982).

The belief merging version of the principle of indifference is the assumption of equal reliability in lack of information about the relative reliability of the sources. In the subjective interpretation of probability, the probability of an event is the degree of belief in that event happening (Hájek 2012); in belief merging, the weight of a source is the likeliness of the formulae it provides being true, or at least close to truth (Cholvy 1998; Darwiche and Marquis 2004; Konieczny et al. 2004; Lin 1996; Revesz 1997). The event “formula F is true in the real world” provides a qualitative connection of probability with merging. The principle of indifference translates into the assumption of equal reliability.

The probability version of sources of unknown reliability is the lack of knowledge of the probability of events. Economists distinguish between risk (known probability) and Knightian uncertainty (unknown probability) (Nishimura and Ozaki 2007). An often-used example is the urn containing twenty yellow balls and forty balls of another color, which may be either blue or green; these forty balls are either all blue or all green, but which of the two is not known. This scenario involves both risk (the probability of yellow or not yellow is known) and Knightian uncertainty (the presence of blue balls is unknown).

This urn suggests a way to deal with the problem in belief merging. The probability of drawing a yellow ball is always one third, but assuming the same for blue and green is as if the urn contained twenty balls for each color. This is acceptable for a single drawn, but not in general. The probability of drawing two balls of the same color (putting the first ball back in the urn) under the assumption of equal probability is \(\frac{1}{3}\) instead of \(\frac{1}{3}\frac{1}{3} + \frac{2}{3}\frac{2}{3}\). The first value is obtained by selecting from the nine possible outcomes of probability \(\frac{1}{9}\) each (random first ball and random second ball) only the three where the balls have the same color: \(3\frac{1}{9} = \frac{1}{3}\). The second value can be obtained by considering the second drawn not independent of the first, but also by calculating the probability under the assumption of forty blue balls: the probability of the two balls being both yellow is \(\frac{1}{3} \frac{1}{3}\), that of being both blue is \(\frac{2}{3} \frac{2}{3}\). Importantly, the very same value is obtained for forty green balls instead. Not only this probability holds in both cases, it resists the addition of information. It holds even if it is later discovered that the urn is made in a factory that normally produces forty green balls, and the blue ball version is a rare collector’s edition.

In terms of belief merging, two sources of unspecified reliability may be equally reliable, or one may be more reliable than the other. All cases are considered, and only what holds in all of them is taken. This is analogous to reasoning from multiple probability distributions (Halpern and Tuttle 1993).

How does this solution work in the example of the three scenarios A, B and C respectively ranked [3, 0], [1, 1] and [0, 3]? Scenario A is preferred if the second source is much more reliable than the first, scenario B if they are equally reliable and scenario C if the first is much more reliable than the second. If reliability information is absent, none of the three scenarios can be considered more likely than the others. If the first source is much more reliable than the second, scenario A is much less likely than the others. More information (the relative reliability of the sources) leads to more information (from all three scenarios to only B and C).

A result in this article is that the disjunction of all maxcons (Ammoura et al. 2015; Baral et al. 1992; Benferhat et al. 1997; Brewka 1989; Dubois et al. 2016; Grant and Hunter 2011; Konieczny and Pino Pérez 2011) is the result of merging formulae of unknown reliability using the drastic distance. This result invalidates the view that maxcons are unsuitable for merging since they do not take into account the distribution of information (Konieczny 2000; Konieczny and Pino Pérez 2011). This may be the case under equal or otherwise specified reliability, but maxcons do exactly what they should when reliability information is absent.

Another result is a motivation for disjunctive merging, the kind that only selects models of the formulae (Everaere et al. 2010a; Liberatore and Schaerf 1998). It results from assuming that one of the sources is completely reliable, but which one is not specified.

Technically, merging is defined by selecting the models at a minimal weighted distance from the formulae provided by the sources. The drastic and the Hamming distances are considered as two relevant examples. After Sect. 2 fixes the formal language used and other notions related to merging, Sect. 3 defines merging when reliability is completely or partially absent; this definition is based on the concept of weighted distance between models and formulae. Sect. 4 shows results about a property of models that makes them relevant to merging. Sect. 5 and Sect. 6 analyze the case of the drastic and the Hamming distance. Sect. 7 shows which postulates are satisfied, while Sect. 8 concentrates on a specific condition of merging. Sect. 9 considers alternative ways of merging: sum of powers, leximax and leximin. Sect. 10 briefly considers the case of sources providing more than one formula. Sect. 11 discusses the results obtained in this article. An appendix contains all proofs of theorems and lemmas.

2 Preliminaries

The formulae in this article are propositional over a finite alphabet. Models are represented by the set of literals they satisfy; for example, \(I=\{a,\lnot b,c\}\) is the model assigning false to b and true to a and c. The notation \(I \models F\) indicates that the model I satisfies the formula F. The same symbol is also used to indicate that all models of a set satisfy the formula; for example, \(\{I,J\} \models F\) tells that both I and J satisfy F. The set of all models that satisfy a formula F is denoted \(\mod (F)\).

The formulae to merge are denoted \(F_1,\ldots ,F_m\); a single other formula \(\mu \) contains all integrity constrains—what is known for certainty. Contrary to most previous studies in belief merging, the formulae to be merged are not assumed equally reliable, nor they are assumed to have a certain relative reliability either. The aim of merging is to draw as many conclusions as possible.

Merging is formalized by a function from \(\mu \) and \(F_1,\ldots ,F_m\) to something that represents information. In this article, this function produces the propositional interpretations that model the scenarios that are considered possible as the result of merging. In other words, the codomain of this function is the set of all sets of propositional models over the given alphabet. As an example, merging \(x \wedge z\) and \(y \wedge \lnot z\) under the integrity constraint \(x \wedge y\) may result in two scenarios considered possible: one where x, y and z are all true and another where x and y are true while z is false; the result of the function is the set of the two models \(\{x,y,z\}\) and \(\{x,y,\lnot z\}\).

Definition 1

A merging operator is a function \(\varDelta \) from propositional formulae \(\mu \) and \(F_1,\ldots ,F_m\) to sets of propositional models that satisfy \(\mu \).

The most general situation is when the reliability of the formulae is absent. Other cases are: the formulae are equally reliable; one is much more reliable than the others; none is so (this case was suggested by a referee). These three cases are the implicit assumptions of respectively the usual definition of merging, of disjunctive merging and of merging by majority. Unless otherwise specified, merging with absent reliability means that no reliability information is present at all, not even qualitatively like in the case of no formula being much more reliable than the others.

Merging is often based on a distance measure between models. This is a function from pairs of models to non-negative integers. If I and J are two models, d(IJ) is a non-negative integer that tells how much they differ. This integer is zero if I and J coincide, otherwise it is greater than zero.

Two intuitive and commonly used distances are the drastic and the Hamming distance. The drastic distance is defined by \(dd(I,J) = 0\) if \(I=J\) and \(dd(I,J)=1\) otherwise. The Hamming distance dh(IJ) is the number of literals assigned different truth values by I and J; for example, \(dh(\{a,\lnot b,c\},\{\lnot a,\lnot b,\lnot c\})=2\), since the two models differ on a and c. Other distances can be defined; they are assumed to satisfy \(d(I,I)=0\) and \(d(I,J)>0\) if \(I \not = J\).

3 Merge by weights

In this article, merging is done by minimizing the weighted distance of the models obeying the integrity constraints from the formulae to be merged. The integrity constraints are denoted \(\mu \), the formulae to be merged \(F_1,\ldots ,F_m\). This is the basic settings for belief merging, where each source provides exactly one formula \(F_i\); the case of multiple formulae is considered in a following section. Formulae \(\mu \) and \(F_1,\ldots ,F_m\) are propositional over a finite alphabet.

Merging is based on the distance between models, denoted by d(IJ). Two intuitive and commonly used distances are the drastic and the Hamming distance, defined in the previous section. Distance extends from models to formulae: regardless of which distance is used, d(IF) is the minimal value of d(IJ) for every \(J \models F\). It further extends from a formula to a list of them: the distance between a model and a list of formulae is the array of integers \(d(I,F_1,\ldots ,F_m) = [d(I,F_1),\ldots ,d(I,F_m)]\).

Merging by weighted distance was the historically first way of integrating formulae coming from sources of different reliability (Revesz 1997). Given a vector of positive integers \(W = [w_1,\ldots ,w_m]\), the weighted distance of a model I from the formulae \(F_1,\ldots ,F_m\) is \(W \cdot d(I,F_1,\ldots ,F_m)\), where the dot stands as usual for the scalar product:

$$\begin{aligned}{}[w_1,\ldots ,w_m] \cdot d(I,F_1,\ldots ,F_m) = \sum _{1 \le i \le m} w_i d(I,F_i) \end{aligned}$$

This product defines a single integer telling the aggregated distance from I to the formulae \(F_i\), weighted by the relative reliability of each as represented by the integer \(w_i\). Merging selects the models satisfying the integrity constraints \(\mu \) that have minimal weighted distance from the formulae.

$$\begin{aligned} \varDelta ^{d,W}_\mu (F_1,\ldots ,F_m) = \{I \models \mu \mid W \cdot d(I,F_1,\ldots ,F_m) \text{ is } \text{ minimal } \} \end{aligned}$$

This function depends on two parameters: a model-to-model distance d and a vector of weights \(W=[w_1,\ldots ,w_m]\).

Fixed weights are used when the relative reliability of the sources is given. Weights \(W=[1,\ldots ,1]\) make the scalar product the same as a sum, and weighted merge the same as the usual operators based on the sum of the drastic and Hamming distances. In the notation by Konieczny and Pino Pérez (2011):

$$\begin{aligned} \varDelta ^{dd,[1,\ldots ,1]}= & {} \varDelta ^{d_D,\varSigma } \\ \varDelta ^{dh,[1,\ldots ,1]}= & {} \varDelta ^{d_h,\varSigma } \end{aligned}$$

The dh distance was first used in belief revision by Dalal (1988); for this reason, it is sometimes called “Dalal distance”. Revesz (1993, 1997) used it with weights for belief merging, followed by Lin (1996) and Lin and Mendelzon (1999). Weights reflect the reliability of the sources: the distance from a formula of high weight affects the total more than the distance from a formula of low weight.

When reliability is absent, all possible weight vectors are considered. The set of all weight vectors of positive integers is the focus of this article:

$$\begin{aligned} W_\exists = \{[w_1,\ldots ,w_m] \mid w_i \in \mathbf{N}, w_i > 0\} \end{aligned}$$

Nevertheless, other sets of weight vectors are considered. In some scenarios a source is correct; the others only provide refining information. For example, a cardiologist, a pulmonologist and an allergist may have contrasting opinions about the state of a patient; if the symptoms are caused by a heart disease then the cardiologist is likely to be right on everything, for example on the reason of the breathing problems, even if that contradicts the pulmonologist and the allergist; their opinion only provide some additional insight. In the same way, if the symptoms are caused by an allergy, the allergist is likely right on everything. The same for the pulmonologist. In these cases, one source is totally correct, but which one is unknown.

$$\begin{aligned} W_a = \{ [a,1,\ldots ,1], [1,a,1,\ldots ,1], \ldots , [1,\ldots ,1,a] \} \end{aligned}$$

The value of a for scenarios like that of the three doctors depends on the maximal possible distance between a model and a formula. For the drastic distance, \(a=m+1\) suffices, where m is the number of formulae to be merged. For the Hamming distance, \(a=nm + 1\), where n is the number of variables. The opposite case is that of no source deemed much more reliable than the others. It can be formalized by bounding all weights by a constant.

Finally, merging with fixed weights falls into this generalization as the set comprising a single vector. For example, equal reliability is captured by:

$$\begin{aligned} W_= = \{[1,\ldots ,1]\} \end{aligned}$$

In all these cases, a set of weights W. represents all possible reliability the sources are considered to have. Three relevant such sets are \(W_\exists \), \(W_a\) or \(W_=\). The set \(W_=\) is for equally reliable sources; \(W_\exists \) is the other extreme: no reliability information on the sources is present. Every \(W \in W.\) is an encoding of the reliability of the sources. All of these are plausible alternatives. Every scenario (every model) that is possible when merging with some \(W \in W.\) is possible when merging with W.:

$$\begin{aligned} \varDelta ^{d,W.}_\mu (F_1,\ldots ,F_m) = \bigcup _{W \in W.} \varDelta ^{d,W}_\mu (F_1,\ldots ,F_m) \end{aligned}$$

Merging on a set of weights generates all models obtained by merging with one of these weights. This is different from obtaining a single ordering on the models as done by Benferhat et al. (2014) to solve the related problem of commensurability that occurs when the sources themselves assess the reliability of the formulae they provide.

4 Dominance

If a model is farther from every formula than another, the latter is always preferred to the former regardless of the weights. The second model dominates the first. Despite the seeming triviality of the concept, a number of relevant results follow:

  • if a model has minimal weighted distance for some weights, it is not strictly dominated by another;

  • if the codomain is binary, the converse also holds: a model that is not strictly dominated by another has minimal weighted distance for some weights;

  • for a ternary codomain, there exist two formulae such that their merge does not include an undominated model;

Dominance could be defined over models with respect to formulae, but is simpler to formalize over vectors of integers. It can then be carried over to the distance vectors of two models.

Definition 2

A vector of integers D dominates another \(D'\), denoted \(D \le D'\), if every element of D is less than or equal to the element of the same index in \(D'\). Strict dominance is the strict part of this ordering: \(D < D'\) holds if both \(D \le D'\) and \(D' \not \le D\) hold.

The dominance between the distance vectors of two models is the same as the weak Pareto dominance used in multi-objective decision making (Giagkiozis and Fleming 2014) when the objectives to minimize are the distances between the models and the formulae.

If the distance vector of a model is strictly dominated by that of another, the first is never minimal regardless of the weights. This fact holds because weights are strictly positive.

Lemma 1

For every distance d, vector of weights \(W \in W_\exists \) and model I of \(\mu \), if \(I \in \varDelta ^{d,W}_\mu (F_1,\ldots ,F_m)\) then \(d(J,F_1,\ldots ,F_m) < d(I,F_1,\ldots ,F_m)\) holds for no model J of \(\mu \).

This result is almost trivial, since merging selects the models that have a minimal value of the sum of the distances, each multiplied by a positive weight. The converse does not hold in general, but does in a relevant case: when \(d(I,F_i)\) can only be 0 or 1, or more generally when the codomain of d has size two.

Lemma 2

If the codomain of the distance function d is a subset of cardinality two of \(\mathbf{N}\), I is a model of \(\mu \), \(d(I,F_1,\ldots ,F_m)\) is not strictly dominated by the vector of distances of any other model of \(\mu \), then there exists W such that \(I \in \varDelta ^{d,W}_\mu (F_1,\ldots ,F_m)\).

The last two lemmas imply that the minimal models with arbitrary weights \(W_\exists \) according to a distance of binary codomain are exactly the models whose distance vectors are not dominated by others. Since dominance is the same as weak Pareto dominance, these minimal models are exactly the Pareto set (Giagkiozis and Fleming 2014). This is therefore the result of merging, but only when reliability information is completely absent and the codomain of the distance function is binary.

Theorem 1

If the codomain of the distance d is a subset of cardinality two of \(\mathbf{N}\), then \(\varDelta ^{d,W_\exists }_\mu (F_1,\ldots ,F_m)\) is the set of all models of \(\mu \) of minimal distance vector according to the dominance ordering.

The next question is whether this condition holds for every fixed-size codomain, or whether a codomain of size three is sufficient for making some undominated model to be excluded from merge. The latter is indeed the case in general. A preliminary lemma will be useful in the sequel.

Lemma 3

If \(\mu \) has three models of distance [3, 0], [2, 2] and [0, 3] from \(F_1\) and \(F_2\), then \(\varDelta ^{d,W_\exists }_\mu (F_1,F_2)\) does not contain the model at distance [2, 2].

This is almost the proof of the claim, but the codomain of the Hamming distance has size unbounded, not three. However, a slightly different distance function fixes the problem.

Theorem 2

For some distance d with codomain of size three, there exists I, \(\mu \) and \(F_1,\ldots ,F_m\) such that \(I \models \mu \) and \(I \not \in \varDelta ^{d,W_\exists }_\mu (F_1,\ldots ,F_m)\), but \(d(J,F_1,\ldots ,F_m) < d(I,F_1,\ldots ,F_m)\) does not hold for any \(J \models \mu \).

5 Drastic distance

Merging with all possible weights and the drastic distance dd generates all models of all maximal subsets of \(F_1,\ldots ,F_m\) that are consistent with \(\mu \). This is proved in three steps:

  • dominance with the drastic distance is the same as the containment of the set of formulae \(F_1,\ldots ,F_m\) satisfied by the models;

  • the models of the maxcons are the models that are minimal according to that containment;

  • therefore, the models of the maxcons are exactly the undominated models; by the results in the previous section, they are the models of minimal weighted distance according to some weights.

Maximal consistent subsets (maxcons) have a general definition over lists of sets of formulae, but what is necessary for this article is only the version with a list of two sets, the first comprising a single consistent formula \(\mu \) and the second \(F_1,\ldots ,F_m\). With this limitation, the (possibly non-maximal) consistent subsets and the maximal consistent subsets are defined as:

$$\begin{aligned}&{\mathrm con_\mu (F_1,\ldots ,F_m) =} \{ S \subseteq \{\mu ,F_1,\ldots ,F_m\} \mid \mu \in S \text{ and } S \not \models \bot \} \\&\mathrm maxcon_\mu (F_1,\ldots ,F_m) {=} \{ S {\in } \mathrm con_\mu (F_1,\ldots ,F_m) \mid \not \exists S' {\in } \mathrm con_\mu (F_1,\ldots ,F_m) ~.~ S {\subset } S' \} \end{aligned}$$

Since \(\mu \) is consistent, these sets cannot be empty. To establish the correspondence between models and maxcons, the subset of formulae satisfied by a model is needed.

Definition 3

The set of formulae satisfied by a model I is denoted \(\mathrm subsat(I,F_1,\ldots ,F_m) = \{ F_i \mid I \models F_i\}\).

The basic brick in the proof construction is that dominance of the drastic distance vectors is the same as containment of the subsets of formulae satisfied by models.

Lemma 4

For every pair of models I and J and every formulae \(F_1,\ldots ,F_m\), the following two conditions are equivalent:

$$\begin{aligned} dd(I,F_1,\ldots ,F_m)\le & {} dd(J,F_1,\ldots ,F_m) \\ \mathrm subsat(J,F_1,\ldots ,F_m)\subseteq & {} \mathrm subsat(I,F_1,\ldots ,F_m) \end{aligned}$$

This lemma links the dominance ordering under dd and the containment of \(\mathrm subsat\). The next links the latter with the maxcons.

Lemma 5

A model I of \(\mu \) satisfies some element of \(\mathrm maxcon_\mu (F_1,\ldots ,F_m)\) if and only if \(\mathrm subsat(I,F_1,\ldots ,F_m) \subset \mathrm subsat(J,F_1,\ldots ,F_m)\) holds for no model J of \(\mu \).

The lemma is the final link of the connection between maxcons and dominance under the drastic distance.

Theorem 3

For every consistent formulae \(\mu , F_1,\ldots ,F_m\), the following equality holds:

$$\begin{aligned} \varDelta ^{dd,W_\exists }_\mu (F_1,\ldots ,F_m) = \bigcup _{S \in \mathrm maxcon_\mu (F_1,\ldots ,F_m)} \mod \left( \wedge S\right) \end{aligned}$$

Maxcons have been long used in belief revision (Fagin et al. 1983; Baral et al. 1992; Benferhat et al. 1997; Konieczny and Pino Pérez 2011; Ammoura et al. 2015; Grant and Hunter 2011; Dubois et al. 2016) and nonmonotonic reasoning (Rescher and Manor 1970; Brewka 1989; Ginsberg 1986). Yet, they are sometimes dismissed as “unsuitable for merging” because they do not take into account the distribution of information among the sources (Konieczny 2000; Konieczny and Pino Pérez 2011). This criticism is grounded in the assumption of equal or given reliability. This theorem blocks it from extending to absent reliability; maxcons are weighted merge with the drastic distance when reliability information is completely absent. Not only they are suitable for merging, they deal with the common situation where the credibility of the sources cannot be assessed.

An example clarifies why. If reliability information is completely absent, a formula \(\lnot x\) provided by two sources cannot beat a formula x provided by one source, since the one source may be much more reliable than the two. Merging by maxcons collects as many formulae as possible while retaining consistency; each maxcon may come from the most reliable sources, making the number of formulae itself irrelevant.

This is not the case when some reliability information is present. A source cannot beat all others when the formulae have comparable reliability, as formalized by weights bounded by a constant. It also cannot when all formulae have the same reliability. This difference supports distinguishing absent reliability from equal reliability.

6 Hamming distance

The Hamming distance dh has a codomain of more than two elements. Therefore, the previous results about binary codomains do not apply. Some existence results are proved:

  • every given set of distance vectors is obtainable from some formulae \(\mu ,F_1,\ldots ,F_m\);

  • for some \(\mu ,F,F'\), merging with all possible weight vectors is not equivalent to merging with subexponentially many weight vectors.

Merging with the Hamming distance does not have a simple equivalent form like for the drastic distance, which selects the models that are not strictly dominated by others. The same does not hold in general: a model that is undominated may still be excluded in the merging. This was proved abstractly by three distance vectors [3, 0], [2, 2] e [0, 3]. With the Hamming distance, these distance vectors can be obtained from concrete formulae. So can every set of distance vectors, actually. This existence result is analogous to a similar one for maxcons (Liberatore 2015, Lemma 4.6).

Lemma 6

Given an arbitrary set of distance vectors of m elements each, all bounded by an integer n, for some formulae \(\mu \) and \(F_1,\ldots ,F_m\) over nm variables the vectors of Hamming distances from the models of \(\mu \) to \(F_1,\ldots ,F_m\) are exactly the given set of distance vectors.

This theorem allows for an easy way of building counterexamples: rather than providing d and \(\mu ,F_1,\ldots ,F_m\) that have a certain property, that property is shown directly on the set of distance vectors. This method was already used to prove that some undominated models are not selected by merging, for some distance. In particular, it shows this being the case for the Hamming distance.

Another application is the proof that exponentially many weight vectors have to be considered when merging. The definition itself requires all weight vectors to be taken into account: a model is selected if and only if it is selected by at least one of the infinitely many weight vectors in \(W_\exists \). The next lemma shows that at least exponentially many have to be considered. It will be later proved that exponentially many suffice.

Lemma 7

There exist three formulae \(\mu ,F,F'\) on an alphabet of six variables such that every \(W_r\) such that \(\varDelta ^{dh,W_r}_\mu (F,F') = \varDelta ^{dh,W_\exists }_\mu (F,F')\) contains at least two weight vectors.

This lemma shows a pair of formulae that requires at least two weight vectors. This technical result has an abstract implication: since each weight vector encodes a specific way to compare the reliability of the formulae, merging with absent reliability cannot be reduced to merging with any specific reliability degree of the formulae. This makes sense, as reliability is not objective, like for example the values of variables, but subjective, since it is the strength of believing that a formula is true. If merging were always possible with a single weight vector, that weight vector could be considered as part of reality rather than beliefs.

The lemma requires two formulae of six variables. The claim actually holds for three variables, but the proof would be ad-hoc rather than simply referring to a previous lemma, and is therefore omitted. The claim does not hold for two variables, as proved by exhaustive analysis on the four possible models.

The construction in the lemma can be replicated over many distinct alphabets of six variables each. Each alphabet doubles the number of necessary weight vectors, leading to exponentiality.

Lemma 8

There exists \(\mu ,F_1,\ldots ,F_m\) such that the size of every \(W_r\) for which \(\varDelta ^{dh,W_r}_\mu (F,F') = \varDelta ^{dh,W_\exists }_\mu (F,F')\) is exponential in the size of the formulae.

This result relies on an unbounded number of formulae to be merged. With two formulae, a number of weight vectors linear in the number of variables suffices.

7 Postulates

Merging with unknown weights depends on the distance function and the set of weight vectors. Some postulates for belief merging hold for all sets of weights vectors (IC0-IC2 and IC7), others only for some including \(W_\exists \) (IC3-IC6), and one does not hold for \(W_\exists \) (IC8). Merging with \(W_\exists \) cannot be expressed as a preorder, not even a partial one.

Postulates IC0-8 (Konieczny and Pino Pérez 2002a) cannot all hold, since a merging operator satisfying all of them can be expressed as a selection of models of \(\mu \) that are minimal according to some total preorder that depends only on \(F_1,\ldots ,F_m\). Actually, not even a partial preorder expresses merging with all possible weight vectors.

Theorem 4

No partial preorder \(\le \) depending on \(F_1\) and \(F_2\) only is such that \(\varDelta ^{dh,W_\exists }_\mu (F_1,F_2) = \min (\mod (\mu ),\le )\).

A consequence of this theorem is that merging with all possible weight vectors \(W_\exists \) does not satisfy all postulates, since that would imply that merging could be expressed by a preorder. Some postulates are not satisfied. Others are.

Some postulates hold for every set of weight vectors, others only for some. Some postulates hold only if the distance function satisfies the triangle inequality, others hold even if d(IF) is not defined in terms of a distance among models d(IJ). The latter requires \(d(I,F) \in \mathbf{N}\) and \(d(I,F)=0\) if and only if \(I \models F\). In the following summary, this case is described as “a model-formula distance”. In this section, E is sometimes used in place of \(F_1,\ldots ,F_m\) following the notation by Konieczny and Pino Pérez (2011). This simplifies some formulae.

  1. IC0

    \(\varDelta ^{d,W.}_\mu (E) \subseteq \mod (\mu )\) holds for every model-formula distance and non-empty set of weight vectors

  2. IC1

    if \(\mu \) is consistent, then \(\varDelta _\mu (E)\) is not empty holds for every model-formula distance and non-empty set of weight vectors

  3. IC2

    if \(\bigwedge E\) is consistent with \(\mu \), then \(\varDelta ^{d,W.}_\mu (E) = \mod (\mu ) \cap \mod (\bigwedge E)\) holds for every model-formula distance and non-empty set of weight vectors

  4. IC3

    if \(E_1 \equiv E_2\) and \(\mu _1 \equiv \mu _2\), then \(\varDelta ^{d,W.}_{\mu _1}(E_1) = \varDelta ^{d,W.}_{\mu _2}(E_2)\) holds for every model-formula distance and non-empty set of weight vectors that contains every permutation of every vector it contains (\(W_\exists \) has this property, as well as \(W_a\) for every \(a \in \mathbf{N}\) with \(a > 0\).

  5. IC4

    if \(F_1 \models \mu \) and \(F_2 \models \mu \) then \(\varDelta ^{d,W.}_\mu (F_1,F_2) \cap \mod (F_1)\) is not empty if and only if \(\varDelta ^{d,W.}_\mu (F_1,F_2) \cap \mod (F_2)\) is not empty. holds if W. contains every permutation of every vector it contains and d satisfies the triangle inequality: \(d(I,K) + d(K,J) \ge d(I,J)\) (both dd and dh have this property); if any of these two conditions do not hold, a counterexample shows that the postulate does not hold

  6. IC5

    \(\varDelta ^{d,W.'}_\mu (F_1,\ldots ,F_k) \cap {} \varDelta ^{d,W.''}_\mu (F_{k+1},\ldots ,F_m) {} \subseteq {} \varDelta ^{d,W.}_\mu (F_1,\ldots ,F_m)\) requires W. to be the Cartesian product of two sets of weight vectors \(W.'\) and \(W.''\) whose vectors have size k and \(m-k\), respectively

  7. IC6

    if \(\varDelta ^{d,W.'}_\mu (F_1,\ldots ,F_k) \cap {} \varDelta ^{d,W.''}_\mu (F_{k+1},\ldots ,F_m)\) is not empty, it contains \(\varDelta ^{d,W.}_\mu (F_1,\ldots ,F_m)\) requires W. to be the Cartesian product of two sets of weight vectors \(W.'\) and \(W.''\) whose vectors have size k and \(m-k\), respectively

  8. IC7

    \(\mod (\mu ') \cap \varDelta ^{d,W.}_\mu (E) \subseteq \varDelta ^{d,W.}_{\mu \wedge \mu '}(E)\) holds for every model-formula distance and non-empty set of weight vectors

  9. IC8

    if \(\mod (\mu ') \cap \varDelta ^{d,W.}_\mu (E)\) is not empty, then \(\varDelta ^{d,W.}_{\mu \wedge \mu '}(E) \subseteq \varDelta ^{d,W.}_\mu (E)\) does not hold for the Hamming distance dh and the set of all weight vectors \(W_\exists \)

The formal proofs of these claims follow. First, postulates IC0, IC1, IC2 and IC7 hold for every non-empty set of weight vectors W. and model-formula distance.

Lemma 9

For every model-formula distance d and non-empty set of weight vectors W., the merging operator \(\varDelta ^{d,W.}\) satisfies postulates IC0, IC1, IC2 and IC7.

Postulate IC3 includes the case where the order of the formulae is changed. This affects the weight vectors: they must be allowed to change their internal order accordingly.

Lemma 10

If W. contains every permutation of every vector it contains, then IC3 holds. For some set of weight vectors that does not include a permutation of one of its elements, IC3 does not hold.

These lemmas do not require d(IF) to be defined in terms of a distance between models d(IJ). The next one does, and additionally needs the triangle inequality.

Lemma 11

If W. contains every permutation of every vector it contains and d satisfies the triangle inequality \(\forall I,J,K . d(I,K) + d(K,J) \ge d(I,J)\), then IC4 holds. For some set of weight vectors that does not include a permutation of one of its elements IC4 does not hold. The same for some distance not satisfying the triangle inequality.

Since \(W_\exists \) is symmetric and both dd and dh satisfy the triangle inequality, Postulate IC4 holds in these two cases. Actually, for \(W_\exists \) the distance does not matter, and \(\varDelta ^{d,W_\exists }_\mu (F_1,\ldots ,F_m)\) always contains some models of every \(F_i\) that is consistent with \(\mu \). If the maximal value of the distance from \(F_1\) and from \(F_2\) is k, the weight vectors \([k+1,1]\) and \([1,k+1]\) suffice. The first guarantees that every model of \(F_1\) is always better than one of \(\lnot F_1\), no matter how close the second is to \(F_2\). The same for the second weight vector.

This lemma shows an effect of the triangle inequality on belief merging. It is a quite natural requirement and is obeyed by both the drastic and the Hamming distance, but is mostly useless in belief merging (Konieczny and Pino Pérez 2011). Besides proving that a certain merging operator satisfies an additional postulate (Konieczny and Pino Pérez 2002a), so far it only seemed to affect the infinite-alphabet case (Chacón and Pino Pérez 2006) and the application of belief revision to case-based reasoning (Cojan and Lieber 2012).

Postulates IC5 and IC6 require special care even to be formulated. Informally, they tell that merging \(F_1,\ldots ,F_k,F_{k+1},\ldots ,F_m\) is the same as merging \(F_1,\ldots ,F_k\), merging \(F_{k+1},\ldots ,F_m\) and then conjoining the two results if they do not conflict. This is simple to express if no weights are involved, otherwise each of these three mergings is defined over its set of weights. If these are unrelated, like \([1,\ldots ,1,1\ldots ,1]\) for the overall merge and \([10,1,\ldots ,1]\) and \([1,\ldots ,10]\) for merging the two parts, the three results cannot be expected to be coherent.

This is why Postulates IC5 and IC6 cannot be said to be obeyed plain and simple. Rather, they are satisfied only when the sets of weights are related in the appropriate way.

Lemma 12

If \(\varDelta ^{d,W.'}_\mu (F_1,\ldots ,F_k) \cap {} \varDelta ^{d,W.''}_\mu (F_{k+1},\ldots ,F_m)\) is not empty, it coincides with \(\varDelta ^{d,W.}_\mu (F_1,\ldots ,F_m)\), where W. is the Cartesian product of \(W.'\) and \(W.''\) (postulates IC5 and IC6).

IC8 does not hold. The following counterexample shows that for the Hamming distance dh and the set of all weight vectors \(W_\exists \).

Theorem 5

There exist \(\mu \), \(\mu '\), \(F_1\) and \(F_2\) such that \(\mod (\mu ') \cap \varDelta ^{dh,W_\exists }_\mu (F_1,F_2)\) is not empty but \(\varDelta ^{dh,W_\exists }_{\mu \wedge \mu '}(F_1,F_2) {} \not \subseteq {} \varDelta ^{dh,W_\exists }_\mu (F_1,F_2)\).

This counterexample completes the analysis of the basic postulates IC0-IC8. Two additional ones exist: majority and arbitration. The first tells that a formula repeated enough times is entailed by the result of merging; the second was initially defined as the irrelevance of the number of repetitions, and has a newer definition that is difficult to summarize in words.

Majority does not hold with \(W_\exists \). Not that it should. No matter how many times a formula is repeated, regardless of how many sources supports it, its negation may come from a single source that is more reliable than all the others together. When reliability is uncertain, this case has to be taken into account. It is not even uncommon in practice: many commonly held belief are in fact false.

Many commonly held beliefs are in fact false: Napoleon was short (Dunan 1963); diamonds had been typical gemstones for engagement rings since a long time (Epstein 1982) the red telephone is a telephone line, and one of its end is in the White House (Clavin 2013); meteorites are always hot when they reach the Earth’s surface; flowering sunflowers turn to follow the sun (only the gems do); the Nazis issued an ultimatum before the Ardeatine massacre (something even witnesses of the time believe) (Mazzoni 2003, p. 155); fans in closed rooms kill people (many people in Korea believed this). A page on Wikipedia lists more than a hundred of commonly believed facts that are in fact false (Wikipedia 2017b). The material was enough for a 26-episodes TV show (Wikipedia 2017a).

A view of belief merging is that it formalizes the process of information aggregation by human agents. The above scenarios indicate that unanimity is often a driving mechanism of believing: hearing and reading many times that Napoleon was short leads to believing he was without questioning. Yet, unanimity is not majority. A single person with a funny haircut on TV may at least cast a doubt.

All of this shows that no matter how many times a fact is repeated, when no reliability information is present, it may still be falsified by a single reliable source. This is what the following theorem formally proves.

Theorem 6

There exists \(F_1,F_2\) such that \(\varDelta ^{d,W_\exists }_\mathsf{true}(F_1,F_2,\ldots ,F_2) \not \subseteq \mod (F_2)\), where \(F_2\) is repeated an arbitrary number of times.

Majority does not hold in the most unconstrained case \(W_\exists \) where weights are arbitrary. This does not mean that majority never applies. It means that it does not apply when no information about the reliability of the sources is present. In many other cases, it applies. When sources are considered equally reliable, it applies. It holds for \(W_= = \{[1,\ldots ,1]\}\), which formalizes exactly this situation. The operator \(\varDelta ^{d,W_=}_\mu (F_1,\ldots ,F_m)\) coincides with the operator \(\varDelta ^{d,\varSigma }_\mu (F_1,\ldots ,F_m)\) defined by Konieczny and Pino Pérez (2011), which satisfies majority (Konieczny and Pino Pérez 2002a).

Where is the boundary between majority and non-majority operators? The majority condition tells that no source is arbitrarily more reliable than the others. Since reliability is formalized by weights, it tells that no weight is arbitrarily large.

Theorem 7

If all weights in the vectors in W. are lower than a constant and d is an arbitrary distance, for every \(\mu \) and \(F_1,\ldots ,F_o,F_{o+1},\ldots ,F_m\), there exists n such that \(\varDelta ^{d,W.}_\mu {} (F_1,\ldots ,F_o,F_{o+1},\ldots ,F_{o+1},\ldots ,F_m,\ldots ,F_m) {\subseteq } \varDelta ^{d,W.}_\mu (F_{o+1},\ldots ,F_m)\), where the formulae from \(F_{o+1}\) to \(F_m\) are repeated n times.

Weights bounded by a constant indicate that no source is ever considered arbitrarily more reliable than the others. This is a relevant case. So it is the case where a source may be much more reliable than the others, as motivated above by the example of the commonly believed facts that are in fact false and previously by the example of the three doctors.

Arbitration was initially defined as the opposite condition of irrelevance of the number of repetitions (Konieczny and Pino Pèrez 1998; Meyer 2001). This property holds for \(W_\exists \). The following theorem proves an equivalent formulation of it.

Lemma 13

For every \(\mu ,F_1,\ldots ,F_m\) it holds:

$$\begin{aligned} \varDelta ^{d,W_\exists }_\mu (F_1,\ldots ,F_m) = \varDelta ^{d,W_\exists }_\mu (F_1,\ldots ,F_m,F_m) \end{aligned}$$

A newer version of the arbitration postulate is expressed in terms of the preorder between models as: if \(I<_{F_1}J\), \(I<_{F_2}J'\) and \(J \equiv _{F_1,F_2} J'\) then \(I <_{F_1,F_2} J\). As proved by Theorem 4, merging with absent reliability cannot be expressed as a preorder, total or otherwise. The expression of the postulate in terms of formulae is even more convoluted, and is not clear whether it makes sense when merging is not expressible in terms of a preorder.

8 One reliable source

The case of one reliable, unspecified source is captured by the set of weights \(W_a\) when the number a is large enough: more than the maximal distance between a model and the formulae. For these weight vectors, a single source may take over all other ones. Such a situation is not unlikely in practice, as exemplified by the scenario of the three doctors: the one specialized in the field of the actual illness is almost certainly right, but the actual illness is debated. Another example is that of the facts commonly held true, where the opinion of a real expert may confute them regardless of how many people believe them.

The assumption of one reliable source gives rise and motivates a condition already in the literature: the disjunctive property. Technically, merging by the weight vectors \(W_a\) with a sufficiently large a ensures the disjunctive property. Conceptually, the disjunctive property formalizes the assumption of one reliable source.

The disjunctive property was defined on two formulae as Postulate 7 by Liberatore and Schaerf (1998) and later generalized to an arbitrary number of formulae with integrity constraints by Everaere et al. (2010a). In terms of models, it has a simple and intuitive expression. Every model is a possible state of the world; merging only selects the ones that at least one of the sources considers possible. In formulae, a model I is in the result of merging only if \(I \models F_i\) for at least one of the merged formulae \(F_i\). Since I must also satisfy the integrity constraints \(\mu \), this requirement is lifted when none of the formulae \(F_i\) is consistent with \(\mu \).

Definition 4

A merging operator \(\varDelta \) is disjunctive if it satisfies the disjunctive property: \(\varDelta _\mu (F_1,\ldots ,F_m) \subseteq \mod (F_1 \vee \cdots \vee F_m)\) holds if at least one of the formulae \(F_1,\ldots ,F_m\) is consistent with \(\mu \).

This condition is not satisfied by \(\varDelta ^{dh,W_=}\). As a result, is not satisfied by \(\varDelta ^{dh,W_\exists }\) either, since \(W_= \subset W_\exists \).

The disjunctive property fails in the case of equal reliability (formalized by \(W_=\)) and completely absent reliability information (formalized by \(W_\exists \)). This is one part of the claim that the disjunctive property is a formalization of the assumption of one reliable source. The other is that it succeeds when one unspecified source is much more reliable than the others, formalized by \(W_a\) with a large value of a.

Definition 5

(Liberatore and Schaerf 1998)

Merging by closest pairs of models is defined from the ordering between pairs of models \(\langle I,J \rangle \le _{dh} \langle I',J' \rangle \) if and only if \(dh(I,J) \le dh(I',J')\) by selecting the models in all minimal pairs:

$$\begin{aligned} F_1 \varDelta _D F_2= & {} \{ I,J \mid \langle I,J \rangle \in \min (\{\langle K,L \rangle \mid K \in \mod (F_1), \\&L \in \mod (F_2)\}, \le _{dh}) \} \end{aligned}$$

This definition is framed in the general framework of merging with the set of weights \(W_{n+1} = \{[1,n+1], [n+1,1]\}\), the specific form of \(W_a\) when a is the number of variables increased by one and merging is between two formulae. Since n is the maximal value of the Hamming distance, this set \(W_a\) characterizes exactly the assumption that one of the two formulae is reliable, but no information about which is present. This assumption leads to merging by closest pairs of models.

Theorem 8

For every pair of satisfiable formulae \(F_1\) and \(F_2\) over an alphabet of n variables, it holds \(F_1 \varDelta _D F_2 = \varDelta ^{dh,W_{n+1}}_\mathsf{true}(F_1,F_2)\).

A disjunctive operator on m formulae is obtained similarly when all formulae are consistent and the integrity constraints are void: \(\mu =\mathsf{true}\).

Theorem 9

For every distance d bounded by k, if \(F_1,\ldots ,F_m\) are satisfiable then \(\varDelta ^{d,W_{k m}}_\mathrm{true}(F_1,\ldots ,F_m)\) is a disjunctive merging operator.

The weight vectors in these theorems provide an alternative view of the disjunctive property. Rather than being a principle by itself, it is a formalization of the assumption that a single source is fully reliable, but which one is not specified. What the other sources tell is kept into account, but not as much as contradicting the reliable source. In terms of formulae, the other formulae help in selecting some of the models of the reliable formula, but do not drive the choice outside the set of these models. The result of this selection is always a group of these models, a subset of the models of the reliable formula. However, which formula is reliable is not specified. It may be every one of them. For each one, merging may only select some of its models. Overall, only the models of the formulae can be in the result of merging.

This mechanism interprets the principle of indifference in belief merging in the right way: rather than assuming that all sources are equally reliable, one of them is taken as completely right, but this is done for each of them in turn. Indifference is realized by symmetry, not equality.

Many interesting operators are not disjunctive (Konieczny and Pino Pérez 2011; Everaere et al. 2010a). An operator may not be disjuncive because it interprets the principle of indifference in a different way, or because it does not follow the principle of indifference at all. Indifference is not a universal rule. It is the formalization of the absence of reliability information. When the credibility of the sources are given, or is believed to be equal or comparable, the principle of indifference does not apply. Forcing it on all operators would be a gross mistake.

9 Other aggregator functions

Merging selects models of minimal sum of weighted distances. The sum was historically the first way of combining distances. Others were later invented. Few of the properties studied in this article change when switching to other mechanisms: sum of powers, leximax and leximin ordering (Konieczny and Pino Pérez 2011). The main differences are that leximax produces all undominated models even if the codomain of the distance function is not binary and that leximin produces the models of maximal-size maxcons instead of all maxcons when using the drastic distance.

9.1 Merging by sum of powers

Instead of adding the weighted distances, merging by sum of powers adds their weighted powers (Konieczny and Pino Pérez 2002b). The power could be the square (power 2) or an arbitrary positive integer n. The ordering between two weighted vector distances is:

$$\begin{aligned}{}[v_1,\ldots ,v_m]<^n [u_1,\ldots ,u_m] \text{ iff } \sum _{1 \le i \le m} w_i v_i^n < \sum _{1 \le i \le m} w_i u_i^n \end{aligned}$$

This order defines merging by a single weight vector:

$$\begin{aligned}&\varDelta ^{d,W,\varSigma ^n}_\mu (F_1,\ldots ,F_m) = \{ I \in \mod (\mu ) \mid \not \exists J \in \mod (\mu ) .\\&W \cdot d(I,F_1,\ldots ,F_m) <^n W \cdot d(J,F_1,\ldots ,F_m) \} \end{aligned}$$

Merging by a set of weight vectors is the union of merging by each:

$$\begin{aligned} \varDelta ^{d,W.,\varSigma ^n}_\mu (F_1,\ldots ,F_m) = \bigcup _{W \in W.} \varDelta ^{d,W,\varSigma ^n}_\mu (F_1,\ldots ,F_m) \end{aligned}$$

Given any distance function d, the function defined by \(d'(I,F_i) = d(I,F_i)^n\) is also a distance function. It is binary if and only if d is binary. Therefore, all results involving arbitrary distance functions or arbitrary binary distance functions carry over from merging by sum to merging by sum of powers:

  1. Lemma 1:

    merging only produces models that are not strictly dominated by others;

  2. Theorem 1:

    if the distance function has binary codomain, merging produces exactly the models that are not strictly dominated by others;

  3. Theorem 3:

    merging by the drastic distance produces the union of the models of all maxcons.

The latter holds because dd is binary, which implies that \(d'(I,F_i) = dd(I,F_i)^n\) is binary as well. Therefore, merging produces exactly the undominated models. This is the same as the result of merging by the drastic distance without powering the distances. The latter is proved by Theorem 3 to be the union of the models of all maxcons.

The results requiring specific values may not carry over when powering the distances. For example, the proof of Theorem 2 involves three models with vector distances [3, 0], [2, 2] and [0, 3], where the second is not dominated by the others but is not in the result of merging. This is not the case when squaring the distances, as [4, 4] is less than [9, 0] and [0, 9]. Yet, this lost property is found in the distance vectors [4, 0], [3, 3] and [0, 4].

Theorem 10

For some distance d with codomain of size three, there exists I, \(\mu \) and \(F_1,\ldots ,F_m\) such that \(I \models \mu \) and \(I \not \in \varDelta ^{d,W_\exists ,\varSigma ^2}_\mu (F_1,\ldots ,F_m)\) hold, but \(d(J,F_1,\ldots ,F_m) < d(I,F_1,\ldots ,F_m)\) does not hold for any \(J \models \mu \).

The preconditions of this theorem are satisfied by the Hamming distance, since this distance produces every possible set of distance vectors by Theorem 6. As a result, merging by squared distances may not generate some models of \(\mu \) that are not dominated by others.

9.2 Leximax merging

The leximax ordering is the lexicographic order between two vectors sorted in descending order (Konieczny and Pino Pérez 2002a). The vectors \(W \cdot d(I,F_1,\ldots ,F_m)\) are sorted so that each element is less than or equal to the previous, and compared according to the lexicographic order. The minimal models form the result of leximax merging with a single weight vector \(\varDelta ^{d,W,leximax}_\mu (F_1,\ldots ,F_m)\). The result of merging with a set of weight vectors \(\varDelta ^{d,W.,leximax}_\mu (F_1,\ldots ,F_m)\) is the union of merging with each.

The dominance ordering is maintained when sorting vectors in descending order. If the first vector is less than or equal to the second, it remains so after sorting both.

Lemma 14

If \(v = [v_1,\ldots ,v_m]\) and \(u = [u_1,\ldots ,u_m]\) are two vectors of integers such that \(v_i \le u_i\) holds for every index i, the same holds for the result of sorting v and u in descending order.

This result only proves that the ordering before sorting implies that after. The following results require the ordering to be strict. This case is covered by the following lemma.

Lemma 15

If \(v = [v_1,\ldots ,v_m]\) and \(u = [u_1,\ldots ,u_m]\) are two vectors of integers such that \(v_i \le u_i\) holds for every index i and \(v_i < u_i\) for some index i, the same holds for the result of sorting v and u in descending order.

This result allows extending Lemma 1 to leximax merging: it does not generate any model dominated by another.

Lemma 16

For every distance d, vector of weights \(W \in W_\exists \) and model I, if \(I \in \varDelta ^{d,W,leximax}_\mu (F_1,\ldots ,F_m)\) then \(d(J,F_1,\ldots ,F_m) < d(I,F_1,\ldots ,F_m)\) holds for no model J of \(\mu \).

Leximax merging generates exactly the undominated models. This is analogous to Lemma 2, but does not require the codomain of the distance function to be binary.

Lemma 17

If I is a model of \(\mu \) and \(d(I,F_1,\ldots ,F_m)\) is not strictly dominated by the vector of distances of any other model of \(\mu \), then there exists W such that \(I \in \varDelta ^{d,W,leximax}_\mu (F_1,\ldots ,F_m)\).

Combining the last two results: leximax merging generates exactly the models of \(\mu \) that are not dominated by others. This links leximax to maxcons.

Theorem 11

For every consistent formulae \(\mu ,F_1,\ldots ,F_m\), the following equality holds:

$$\begin{aligned} \varDelta ^{d,W_\exists ,leximax}_\mu (F_1,\ldots ,F_m) = \bigcup _{S \in \mathrm maxcon_\mu (F_1,\ldots ,F_m)} \mod \left( \wedge S\right) \end{aligned}$$

9.3 Leximin merging

The leximin ordering is similar to the leximax ordering, but sorts vectors in ascending order instead of descending (Everaere et al. 2010a). Leximin merging generates the minimal weighted vectors \(\varDelta ^{d,W,leximin}_\mu (F_1,\ldots ,F_m)\) according to this ordering. The union of these for all \(W \in W.\) defines \(\varDelta ^{d,W.,leximin}_\mu (F_1,\ldots ,F_m)\).

A model dominated by another is not selected by leximin merging. Its distance vector is greater than that of the other. This property weathers multiplying each distance by the same weight. It also weathers sorting the two vectors in descending order by Lemma 15. It again weathers inverting the order of the two vectors. Therefore, the dominated model is greater than another in the leximin order, and is therefore not selected by leximin merging.

The converse is however not the case even if the distance function has binary codomain. Yet, maxcons are still related to leximin merging with the drastic distance.

The counterexample is based on three models of distance vectors [1, 1, 0, 0], [0, 0, 1, 0] and [0, 0, 0, 1]. Multiplying these vectors by the weights \([w_1,w_2,w_3,w_4]\) results in \([w_1,w_2,0,0]\), \([0,0,w_3,0]\) and \([0,0,0,w_4]\). Since all weights are larger than zero, ordering these vectors produces \([0,0,w_1,w_2]\), \([0,0,0,w_3]\) and \([0,0,0,w_4]\) if \(w_1 \le w_2\), otherwise \([0,0,w_2,w_1]\), \([0,0,0,w_3]\) and \([0,0,0,w_4]\). Regardless, the first vector is not minimal because both \(w_1\) and \(w_2\) are larger than zero.

Model selection is primarily based on the length of the initial string of zeros. The drastic distance gives zero when the model satisfies the formula. Therefore, the number of zeros is the number of satisfied formulae. The selected models are therefore those of the largest-size maxcons.

Theorem 12

For every consistent formulae \(\mu ,F_1,\ldots ,F_m\), the following equality holds

$$\begin{aligned} \varDelta ^{dd,W_\exists ,lexmin}_\mu (F_1,\ldots ,F_m) = \bigcup _{S \in \mathrm cardcons_\mu (F_1,\ldots ,F_m)} \mod \left( \wedge S\right) \end{aligned}$$

where

$$\begin{aligned}&{\mathrm cardcons_\mu (F_1,\ldots ,F_m)}{=} \{ S {\in } \mathrm maxcon_\mu (F_1,\ldots ,F_m) \mid \not \exists S' {\in } \mathrm maxcon_\mu (F_1,\ldots ,F_m) ~.~ |S| {<} |S'| \} \end{aligned}$$

This theorem implies that leximin merging does not generate all models of all maxcons. Therefore, it may not produce all undominated models.

10 Sources providing multiple formulae

The previous sections are about combining a number of independent formulae. This is the basic problem of belief merging: each formula comes from a different source; therefore, their reliabilities are independent. This is formalized by the weights being unconstrained in the set \(W_\exists \).

When a source provides more than one formula, each of them is as reliable as its source. The same mechanism employing the weighted sum of the drastic or Hamming distance can be used, but the weights are associated to the sources rather than to the formulae. All formulae from the same source have the same reliability and therefore the same weight. This condition is close in spirit to the unit partitions by Booth and Hunter (2018).

Technically, each source is represented by a set of formulae \(S_i\). Its reliability is encoded by a positive integer \(w_i\). Given a set \(\{S_1,\ldots ,S_m\}\) of such sources, merging is done by selecting the minimal models of the integrity constraints \(\mu \) according to this evaluation:

$$\begin{aligned} v(I) = \sum _{S_i} w_i \sum _{F_i \in S_i} d(I,F_i) \end{aligned}$$

This is the \(\mathrm {DA}^2\) operator (Konieczny et al. 2004) with the sum as intra-source aggregation and the weighted sum as the inter-source aggregation.

The sum is subject to the problem of manipulation: a source may provide the same formula multiple times in order to influence the final result (Chopra et al. 2006); this is a problem especially when merging preferences, but not when merging beliefs from sources of unspecified reliability. Even if a source provides the same formula a thousand times, one of the considered alternatives is that the weight of this source is a thousand times smaller than the others, making such a manipulation ineffective.

The only technical result in this section is that merging with the drastic distance is not the same as the union of the models of the maxcons. This is proved by the following sources with \(\mu =\mathsf{true}\).

$$\begin{aligned} S_1= & {} \{x,y,z\} \\ S_2= & {} \{\lnot x,\lnot y\} \\ S_3= & {} \{\lnot x,\lnot z\} \end{aligned}$$

One of the maxcons of \(\{x,y,z,\lnot x,\lnot y,\lnot z\}\) is \(\{x,\lnot y,\lnot z\}\), which is not obtained when merging with all possible weights. Intuitively, to include the formula x from \(S_1\) in the result, that formula needs to count at least twice as much as each formula \(\lnot x\) from \(S_2\) and \(S_3\), but this implies the same for y and z, which excludes \(\lnot y\) and \(\lnot z\).

Formally, let the weights of the sources be \(w_1,w_2,w_3\). The weighted distances of some relevant models are:

$$\begin{aligned} I = \{x,\lnot y,\lnot z\}&~ v(I) = w_1 2 + w_2 1 + w_3 1 = w_1 2 + w_2 + w_3&\\ J = \{\lnot x,\lnot y,\lnot z\}&~ v(J) = w_1 3 + w_2 0 + w_3 0 = w_1 3&\\ K = \{x,y,z\}&~ v(K) = w_1 0 + w_2 2 + w_3 2 = w_2 2 + w_3 2&\end{aligned}$$

In order for I to be minimal, v(I) must be less than or equal to v(J) and v(K):

$$\begin{aligned} w_1 2 + w_2 + w_3\le & {} w_1 3 \\ w_1 2 + w_2 + w_3\le & {} w_2 2 + w_3 2 \end{aligned}$$

This system of inequalities is infeasible. The first implies \(w_2 + w_3 \le w_1\), which makes the right-hand side of the second become less than or equal to \(w_1 2\), while it should instead be greater than or equal to \(w_1 2 + w_2 + w_3\), and therefore greater than \(w_1 2\). This proves that \(\{x,\lnot y,\lnot z\}\) is not a minimal model for any weight vector.

A similar example shows that merging does not generate the models of the conjunctions of the formulae of each source. Let \(S_1=\{x,y\}\) and \(S_2=\{\lnot x,\lnot y\}\). The only maxcons of \(\{\wedge S_1, \wedge S_2\}\) are \(x \wedge y\) and \(\lnot x \wedge \lnot y\), but merging with all possible weights selects all four models over x and y, since this is the result when \(W=[1,1]\).

11 Conclusions

Sometimes the information to be merged comes from sources of equal reliability. In such cases, merging with equal weights is correct. But when no reliability information on the sources is present, assuming weights equal is unwarranted. The difference is not only conceptual but also technical. Theorem 4 shows that merging with absent reliability cannot in general be reduced to a preorder among models, not even a partial one.

A result emerged in the study of this setting is a motivation for merging by maxcons (Baral et al. 1992). This mechanism has sometimes been considered unsuitable for merging because it disregards the distribution of information among sources (Konieczny 2000; Konieczny and Pino Pérez 2011). Such a distribution is important when the sources have the same reliability, or more generally their reliability is given. It is not when reliability information is absent. Theorem 3 shows that merging with maxcons is the same as merging with the drastic distance in absence of any reliability assessment on the sources. The number of repetitions of a formula is irrelevant to this kind of merging—as it should. A formula only occurring once may come from a very reliable source, while its negation is supported only by unreliable sources. Without any assumption on the reliability of the sources, this is a situation to take into account. At the same time, Theorem 3 is limited to merging by the drastic distance. It does not apply when a finer distance would provide a more informative result of merging.

This article not only backs merging by maxcons, but more generally merging by the (MI) postulate: the number of repetitions of a formula is irrelevant to merging. Of course, there are many cases where this postulate should not hold; when their reliability is equal, two sources providing a formula give twice the support for it; the same if no source is arbitrarily more reliable than the others, as formalized by weights bounded by a constant. But when reliability information is completely absent, every numeric evaluation is irrelevant, including doubling the support for a formula like in this case. As already discussed by Meyer (2001), Postulate (MI) may sometimes be right; it is inconsistent with the other postulates IC0-IC8, but the fault is on them. While Meyer blames Postulate 4 of merging without integrity constraints (Konieczny and Pino Pèrez 1998), merging with absent reliability conflicts with IC8.

The most significant outcome of this article is a motivation of choices made in the past, uncovering the implicit assumptions they are based on: maxcons come from completely absent reliability information; the disjunctive property comes from the assumption that one formula is totally reliable, but which one is not specified.

A minor technical contribution of this article is a case for the triangle inequality of the distance function. This property had only a couple of applications in belief revision and merging so far (Chacón and Pino Pérez 2006; Cojan and Lieber 2012; Konieczny and Pino Pérez 2002a), but is generally not required (Konieczny and Pino Pérez 2011). The new consequence of it shown in this article is that it allows satisfying Postulate IC4 when merging in absence of reliability information.

Most results in this article are on merging based on the weighted distances of models from the formulae to merge. The overall picture does not change much when switching to other ways of combining distances like the sum of powers and the leximax and the leximin ordering.

A comparison with related work follows.

The weighted distance from a set of formulae was first used for merging by Revesz (1997), and investigated by Lin (1996). Lin and Mendelzon (1999) and Konieczny and Pino Pèrez (1998) used the unweighted sum. These articles assume either equal or fixed weights, not varying weights like this one.

Benferhat et al. (2014) consider the related problem of commensurability: when the sources themselves assess the reliability of the formulae they provide, they may not use the same scale; this is related to a similar issue in social choice theory. Their study and the present one differ in formalism (ranked bases instead of formulae with distance functions), but they share the principle of considering a set of alternative reliability assessments. There is however an early point of departure: Benferhat et al. 2014 distill a single preorder and then select the models that are minimal according to it; Theorem 4 shows that the same cannot be done in the settings of the present article. A point of contact is the case of the drastic distance: Theorem 1 could be alternatively proved from a result by Benferhat et al. (2014, Propositions 1,2,8).

That reliability information may be partially absent is mentioned by Konieczny (2004) as the starting assumption of his model of belief merging as a game: “The hypothesis for those operators is that all the sources are a priori reliable, or that we know that some sources are less reliable than the others, but without knowing which ones.” The approach taken is however very different, as it proceeds by iteratively assessing the reliability of the sources based on the others.

A related question is what changes between the synthetic and epistemic view of belief merging (Everaere et al. 2010b). When attempting to establish the truth (the epistemic view of belief merging), a single opinion from an expert may take over many other ones. When forming a unified opinion of a group (the synthetic view), deciding by majority may look the only way to proceed. As a matter of facts, majority influence research (Gardikiotis 2011) shows otherwise. A minority view may end up prevailing. An example are trial juries, where the opinion of few jurors sometimes forms the final judgment.

Accepting what is true according to all possible relative reliabilities is analogous to drawing the consequences that hold in all probability measures in a set (Halpern and Tuttle 1993), and can be seen as the formal logic version of the “worst scenario” in economics: “the firm may not be certain about the “relative plausibility” of these boom probabilities. [...] if the firm acts in accordance with certain sensible axioms, then its behavior can be characterized as being uncertainty-averse: when the firm evaluates its position, it will use a probability corresponding to the “worst” scenario” (Nishimura and Ozaki 2007). Belief revision and merging aim at the most knowledge that can be justifiably and consistently obtained; therefore, minimal knowledge takes the place of the least profit, and the worst scenario for a formula is one where it is false.