Keywords

1 Introduction

In real-world applications, constraints may be modeled in conjunctive normal form (CNF), but many tasks relevant in AI and reasoning, such as checks for consistency, validity, clausal entailment, and implicants, can not be executed efficiently on them [9]. Tackling these and other computationally expensive problems is the aim of the knowledge compilation paradigm [13]. The idea is to translate a formula into a language in which the task of interest can be executed efficiently [22]. The knowledge compilation map [22] contains an in-depth discussion of such languages and their properties, and other (families of) languages have been introduced since its publication [21, 25, 29]. The focus in this work is on the language deterministic decomposable negation normal form (d-DNNF) [19]. It has been applied in planning [2, 39], Bayesian reasoning [15], diagnosis [3, 43], and machine learning [28] as well as in functional E-MAJSAT [40], to mention a few, and was also studied from a theoretical perspective [7, 8, 10]. Several d-DNNF compilers are available [20, 30, 37, 48], as well as a d-DNNF reasonerFootnote 1.

Translating a formula from CNF to d-DNNF requires to process the search space exhaustively. The number of variable assignments which need to be checked is exponential in the number of variables occurring in the formula and testing them one by one is out of question from a computational complexity point of view. However, if the formula can be partitioned into subformulae defined over pairwise disjoint sets of variables, these subformulae can be processed independently and the results combined [4]. This may reduce the amount of work per computation significantly. Consider \(F = (a\vee b)\wedge (c\vee d)\) defined over the set of variables . Its search space consists of \(2^4 = 16\) variable assignments. The formula \(F\) can be partitioned into \(F_1 = (a\vee b)\) and \(F_2 = (c\vee d)\) defined over the sets of variables and , respectively, and such that \(F = F_1\wedge F_2\). Due to \(V_1 \cap V_2 = \emptyset \), d-DNNF representations of \(F_1\) and \(F_2\) can be computed independently and conjoined obtaining a d-DNNF representation of \(F\). Moreover, in each computation we only need to check \(2^2 = 4\) assignments. The subformulae \(F_1\) and \(F_2\) are called components due to the original motivation originating in graph theory, and the partitioning process is referred to as decomposition or component analysis. This approach, also called component-based reasoning, is realized in various exact \(\#\)SAT solvers [1, 4, 11, 12, 41, 42, 47], and its success suggests that formulae stemming from real-world applications decompose well enough to generate a substantial amount of work saving.

The formula \(F\) in our example satisfies decomposability [22], i.e., for each conjunction, the conjuncts are defined over pairwise disjoint sets of variables. We call such a formula decomposable. Negations occur only in front of literals, hence it is in decomposable negation normal form (DNNF) [17, 18]. A formula in which for each disjunction its disjuncts are pairwise logically contradictory satisfies determinism [22], i.e., for each disjunction \(C_1 \vee \ldots \vee C_n\) it holds that \(C_i \wedge C_j \equiv \bot \) for and \(i \ne j\). A deterministic DNNF formula is said to be in d-DNNF. Determinism is also met by the language disjoint sum of products (DSOP), which is a disjunction of pairwise contradictory conjunctions of literals, and which is relevant in circuit design [5]. In a previous work [34], we introduced an approach for translating a CNF formula into DSOP based on CDCL with chronological backtracking. The motivation for using chronological backtracking is twofold. First, it has shown not to significantly harm solver performance [33, 38]. Second, pairwise disjoint models are detected without the need for blocking clauses commonly used in model enumeration based on CDCL with non-chronological backtracking. Blocking clauses rule out already found models, but they also slow down the solver, and avoiding their usage in model enumeration by means of CDCL with chronological backtracking has empirically shown to be effective [46]. Enhancing our former approach [34] by component-based reasoning enables us to compute a d-DNNF representation of a CNF formula. Reconsider our previous example, and suppose we obtained and . Now , hence , which is in d-DNNF.

Our Contributions. We present Abstract CNF2dDNNF, ACD for short, a declarative formal framework describing the compilation of CNF into d-DNNF and a proof of its correctness. This abstract presentation allows for a thorough understanding of our method at a conceptual level and of its correctness. If our framework is sound, every implementation which can be modeled by it is sound as well. This comprises optimizations and implementation details, such as caches. ACD combines component-based reasoning and CNF-to-DSOP compilation based on conflict-driven clause learning (CDCL) with chronological backtracking. Disjunctions with pairwise contradictory disjuncts are introduced by decisions and subsequently flipping their value upon backtracking, while conjunctions whose conjuncts share no variable are introduced by unit propagation and decomposition. For the sake of simplicity, in our calculus formulae are partitioned into two subformulae. However, lifting it to an arbitrary number of subcomponents is straightforward, and a corresponding generalization is presented.

2 Preliminaries

Let \(V\) be a set of propositional variables defined over the set of Boolean constants \(\bot \) (false) and . A literal is either a variable \(v \in V\) or its negation \(\lnot {v}\). We refer to the variable of a literal \(\ell \) by \(\textsf {var}(\ell ) \) and extend this notation to sets and sequences of literals and formulae. We consider formulae in conjunctive normal form (CNF) which are conjunctions of clauses which are disjunctions of literals. A formula in disjoint sum of products (DSOP) is a disjunction of pairwise contradictory cubes, which are conjunctions of literals. Our target language is deterministic decomposable negation normal form (d-DNNF), whose formulae are built of literals, conjunctions sharing no variables, and disjunctions whose disjuncts are pairwise contradictory. We might interpret formulae as sets of clauses and cubes and clauses and cubes as sets of literals by writing \(C \in F\) and \(\ell \in C\) to refer to a clause \(C\) in a formula \(F\) and a literal \(\ell \) contained in a clause or cube \(C\), respectively. The empty CNF formula and the empty cube are denoted by \(\top \) and the empty DSOP formula and the empty clause by \(\bot \).

A total variable assignment is a mapping \(\sigma :V \mapsto \mathbb {B} \), and a trail \(I = \ell _1{\ldots }{\ell _n}\) is a non-contradictory sequence of literals which might also be interpreted as a (possibly partial) assignment, such that \(I(\ell ) = \top \) iff \(\ell \in I\). Similarly, \(I(C)\) and \(I(F)\) are defined. We might interpret a trail \(I\) as a set of literals and write \(\ell \in I\) to refer to the literal \(\ell \) on \(I\). The empty trail is denoted by \(\varepsilon \) and the set of variables of the literals on \(I\) by \(\textsf {var}(I) \). Trails and literals can be concatenated, written \( I{J}\) and \( I{\ell }\), given \(\textsf {var}(I) \cap \textsf {var}(J) = \emptyset \) and \(\textsf {var}(I) \cap \textsf {var}(\ell ) = \emptyset \). The position of \(\ell \) on the trail \(I\) is denoted by \(\tau ({I}, {\ell }) \). The decision literals on \(I\) are annotated by a superscript, e.g., \({\ell }^d\), denoting open “left” branches in the sense of the Davis-Putnam-Logemann-Loveland (DPLL) algorithm [23, 24]. Flipping the value of a decision literal can be seen as closing the corresponding left branch and starting a “right” branch, where the decision literal \({\ell }^d\) becomes a flipped literal \(\lnot {\ell }\).

The residual of \(F\) under \(I\), written \({F}|{}_{I} \), is obtained by assigning the variables in \(F\) their truth value and by propagating truth values through Boolean connectives. The notion of residual is extended to clauses and literals. A unit clause is a clause \(\{\ell \}\) containing one single literal \(\ell \). By \(\textsf {units}(F) \) (\(\textsf {units}({F}|{}_{I}) \)) we denote the set of unit literals in \(F\) (\({F}|{}_{I} \)). Similarly, \(\textsf {decs}{(I)} \) denotes the set of decision literals on \(I\). By writing \(\ell \in \textsf {decs}{(I)} \) (\(\ell \in \textsf {units}(F) \), \(\ell \in \textsf {units}({F}|{}_{I}) \)), we refer to a decision literal \(\ell \) on \(I\) (unit literal in \(F\), \({F}|{}_{I} \)). A trail \(I\) falsifies \(F\), if \(I(F) \equiv \bot \), i.e., \({F}|{}_{I} = \bot \). It satisfies \(F\), \(I \models F\), if \(I(F) \equiv \top \), i.e., \({F}|{}_{I} = \top \), and is then called a model of \(F\). If \(\textsf {var}(I) = V\), \(I\) is a total model, otherwise it is a partial model.

The trail is partitioned into decision levels, starting with a decision literal and extending until the literal preceding the next decision. The decision level function returns the decision level of a variable \(v \in V\). If \(v\) is unassigned, \(\delta (v) = \infty \), and \(\delta \) is updated whenever a variable is assigned or unassigned, e.g., \(\delta [v \mapsto d]\) if \(v\) is assigned to decision level \(d\). We define \(\delta (\ell ) = \delta (\textsf {var}(\ell ))\), \(\delta (C) = \textsf {max}\{\delta (\ell ) \mid \ell \in C\}\) for \(C \ne \bot \) and \(\delta (I) = \textsf {max}\{\delta (\ell ) \mid \ell \in I\)} for \(I \ne \varepsilon \) extending this notation to sets of literals. Finally, we define \(\delta (\bot ) = \delta (\varepsilon ) = \infty \). By writing \(\delta [I \mapsto \infty ]\), all literals on the trail \(I\) are unassigned. The decision level function is left-associative, i.e., \(\delta [I \mapsto \infty ][\ell \mapsto d]\) expresses that first all literals on \(I\) are unassigned and then literal \(\ell \) is assigned to decision level \(d\).

Unlike in CDCL with non-chronological backtracking [36, 44, 45], in chronological CDCL [33, 38] literals may not be ordered on the trail in ascending order with respect to their decision level. We write \({I}_{\leqslant n}\) (\({I}_{< n}\), \({I}_{= n}\)) for the subsequence of \(I\) containing all literals \(\ell \) with \(\delta (\ell ) \le n\) (\(\delta (\ell ) < n\), \(\delta (\ell ) = n\)). The pending search space of \(I\) is given by the assignments not yet tested [34], i.e., \(I\) and its open right branches \(R(I)\), and is defined as \(O(I) = I \vee R(I), \text {where } R(I) = \bigvee _{\ell \in \textsf {decs}{(I)}} R_{= \delta (\ell )}(I) \text{ and } R_{= \delta {(\ell )}}(I) = \lnot {\ell } \wedge {I}_{< \delta {(\ell )}} \text { for } \ell \in \textsf {decs}{(I)} \). As an example, for \(I = a{{b}^d}{c}{d}{{e}^d{f}}\), \(O(I) = (a){b}{c}{d}{e}{f}\vee (\lnot {b}){a} { (\lnot {e}){a}{b}{c}{d}}\). Similarly, the pending models of \(F\) are the satisfying assignments of \(F\) not yet detected and which are given by \(F \wedge O(I)\).

3 Chronological CDCL for CNF-to-d-DNNF Compilation

In static component analysis the component structure is computed once, typically as a preprocessing step, and not altered during the further execution. In contrast, in our approach the component structure is computed iteratively adopting dynamic component analysis. Algorithm 1 provides a general schema in pseudo-code. It is formulated recursively, capturing the recursive nature of dynamic component analysis. Lines 1–7 and 11 describe model enumeration based on chronological CDCL [34], while lines 8–10 capture component analysis.

figure k

Now assume unit propagation has been carried out until completion, no conflict has occurred and there are still unassigned variables (line 8). If \({F}|{}_{I} \) can be decomposed into two formulae \(G\) and \(H\), we call CNF2dDNNF recursively on \(G\) and \(H\), conjoin the outcomes of these computations with \(I\) and add the result to \(M\) (line 9). If \(I\) contains no decisions, the search space has been explored exhaustively, otherwise chronological backtracking occurs (lines 10). The working of our approach is shown by an example.

Example 1

Let be a set of propositional variables and be a formula defined over \(V\). The execution is depicted as a tree in Fig. 1. For the sake of readability, we show only the formula on which a rule is executed, represented by a box annotated with its component level. Black arrows correspond to “downward” rule applications, while violet (gray) arrows represent “upwards” rule applications and are annotated with the formula returned by the computation of a component. Ignore the rule names for now, they are intended to clarify the working of our calculus which is presented in Sect. 4. We see that, first, \(a\) is propagated, denoted by the black vertical arrow annotated with \(a\) and the name of the applied rule (Unit). The residual of \(F\) under \(a\) is \({F}|{}_{a} = (\lnot {b}\vee c){d}\wedge (\lnot {b}\vee e){f} {(b\vee \lnot {c}){e}} {(b\vee d){f}} {(g\vee h)}\) (not shown). It contains no unit clause but can be decomposed into \((\lnot {b}\vee c){d}\wedge (\lnot {b}\vee e){f} {(b\vee \lnot {c}){e}} {(b\vee d){f}}\) and \((g\vee h)\wedge \). Two new (sub)components are created (by applying rule Decompose) with component level \(0{1}\) and \(0{2}\), respectively, represented by the shadowed boxes.

Fig. 1.
figure 1

Component structure of \(F\) created by ACD.

Since \((g \vee h)\) can not be decomposed further, model enumeration with chronological CDCL is executed on it (not shown) by deciding \(g\) (rule Decide) satisfying \((g \vee h)\), followed by backtracking chronologically (BackTrue), which amounts to negating the value of the most recent decision \(g\), and propagating \(h\) (Unit). The processing of \((g \vee h)\) terminates with \(g \vee \lnot {g} \wedge h\) (CompTrue, not shown). But before this result can be used further, the subcomponent at component level \(0{1}\) needs to be processed. Its formula is \(G = (\lnot {b}\vee c){d}\wedge (\lnot {b}\vee e){f} {(b\vee \lnot {c}){e}} {(b\vee d){f}}\). It neither contains a unit nor can it be decomposed, hence we take a decision, let’s say, \({b}^d\). Now \({G}|{}_{b} = (c\vee d)\wedge (e\vee f) \), which is decomposed into two components with one clause each and component level \(0{1}{1}\) and \(0{1}{2}\), respectively (Decompose). These formulae can not be decomposed further, and they are processed independently, similarly to \((g \vee h)\). Before \(G\) was decomposed, a decision was taken, and we backtrack combining the results of its subcomponents (ComposeBack). We have \({G}|{}_{\lnot {b}} = (\lnot {c}\vee e)\wedge (d\vee f) \) resulting in two components with component levels \(0{1}{1}\) and \(0{1}{2}\), respectively. They are processed and their results combined, after which the results of the subcomponents of the root component are conjoined with \(a\). There is no decision on the trail, and the process terminates with (ComposeEnd). Notice that although component levels can occur multiple times throughout the computation, they are unique at any point in time.

4 Calculus

Due to its recursive nature, combining the results computed for subcomponents in CNF2dDNNF is straightforward. For its formalization, however, a non-recursive approach turned out to be better suited. Consequently, a method is needed for matching subcomponents and their parent. For this purpose, a component level is associated with each component. It is defined as a string of numbers in \(\mathbb {N}\) as follows. Suppose a component \(\mathcal {C}\) is assigned level “\(d\)” and assume its formula is decomposed into two subformulae. The corresponding subcomponents \(\mathcal {C}_G\) and \(\mathcal {C}_H\) are assigned component levels “\(d{1}\)” and “\(d{2}\)”, respectively, with “\(\cdot \)” denoting string composition. Accordingly, the component level of their parent \(\mathcal {C}\) is given by the substring consisting of all but the last element of their level, i.e., “\(d\)”.Footnote 2 The root component holds the input formula, it has no parent and its component level is zero. A component is closed if no rule can be applied to it, and decomposed if either at least one of its subcomponents is not closed or both its subcomponents are closed, but their results are not yet combined. Components which are neither closed nor decomposed are open.Footnote 3 Closed components may be discarded as soon as their results are combined, and the computation stops as soon as the root component is closed. With these remarks, we are ready to present our calculus.

We describe our algorithm in terms of a state transition system Abstract CNF2dDNNF, ACD for short, over a set of global states \(S\), a transition relation \(\leadsto \,\,\subseteq S \times S\) and an initial global state \(\mathcal {S}_0\). A global state is a set of components. A component \(\mathcal {C}\) is described as a seven-tuple \((F,\,V,\,d,\,e,\,I,\,M,\,\delta )^s\), where \(s \) denotes its component state. It is \(c \) if \(\mathcal {C}\) is closed, \(f \) if \(F\) is decomposed, and \(o \) if \(\mathcal {C}\) is open. The first two elements \(F\) and \(V\) refer to a formula and its set of variables, respectively. The third element \(d\) denotes the component level of \(\mathcal {C}\). If \(d \ne 0\), then , where \(d'\) is the component level of the parent component of \(\mathcal {C}\), as explained above. In this manner, the component level keeps track of the decomposition structure of \(F\) and is used to match parent components and their subcomponents. The number of subcomponents of \(\mathcal {C}\) is given by \(e\), while \(I\) and \(\delta \) refer to a trail ranging over variables in \(V\) and a decision level function with domain \(V\), respectively. Finally, \(M\) is a formula in d-DNNF representing the models of \(F\) found so far. A component is initialized by \((F,\,V,\,d,\,0,\,\varepsilon ,\,\bot ,\,\infty )^o\) and closed after its computation has terminated, i.e., \((F,\,V,\,d,\,0,\,I,\,M,\,\delta )^c\). Notice that in these cases \(e = 0\). The initial global state consists of the root component \(\mathcal {C}_0 = (F,\,V,\,0,\,0,\,\varepsilon ,\,\bot ,\,\infty )^o\) with \(F\) and \(V\) denoting the input formula and \(V = \textsf {var}(F) \), while the final global state is given by where \(M \equiv F\) is in d-DNNF. The transition relation \(\leadsto \) is defined as the union of transition relations \(\leadsto _{\textsf {\textsf {R}}}\), where \(\textsf {R}\) is either Unit, Decide, BackTrue, BackFalse, CompTrue, CompFalse, Decompose, ComposeBack or ComposeEnd. Our calculus contains three types of rules, which can abstractly be described as follows:

figure r

In this description, \(\mathcal {S}\) refers to the subset of the current global state consisting of all components which are not touched by rule R, with \(\uplus \) denoting the disjoint set union, e.g., in \(\alpha \), \(\mathcal {C}, \mathcal {C'} \not \in \mathcal {S}\). An \(\alpha \) rule affects a component \(\mathcal {C}\) turning it into \(\mathcal {C'}\). The rules Unit, Decide, BackTrue, BackFalse, CompTrue, and CompFalse are \(\alpha \) rules. A \(\beta \) rule modifies \(\mathcal {C}\) obtaining \(\mathcal {C}'\) and creates two new components \(\mathcal {C}_1\) and \(\mathcal {C}_2\). Rule Decompose is the only \(\beta \) rule. Finally, a \(\gamma \) rule removes the two components \(\mathcal {C}_1\) and \(\mathcal {C}_2\) from the global state and modifies their parent \(\mathcal {C}\). Rules ComposeBack and ComposeEnd are \(\gamma \) rules. The rules are listed in Fig. 2.

Fig. 2.
figure 2

ACD transition rules.

Model Computation. Rules Unit, Decide, BackTrue, BackFalse, CompTrue, and CompFalse execute model enumeration with chronological CDCL [34] and are applicable exclusively to open components. Unit literals are assigned the decision level of their reason, which might be lower than the current decision level (rule Unit). Decisions can be taken only if the processed formula is not decomposable (Decide). Backtracking occurs chronologically, i.e., to the second highest decision level on the trail, after finding a model (BackTrue) and to the decision level preceding the conflict level after conflict analysis (BackFalse), respectively. In the latter case, the propagated literal is assigned the lowest level at which the learned clause becomes unit and to which a SAT solver implementing CDCL with non-chronological backtracking would backtrack to. Since the literals might not be ordered on the trail in ascending order with respect to their decision level, a non-contiguous part of it is discarded. Finally, a component is closed if its trail contains no decisions and either satisfies its formula (CompTrue) or a conflict occurs at decision level zero, i.e., the conflicting clause has decision level zero (CompFalse). In the former case, the newly found model is recorded.

Component Analysis. Rules Decompose, ComposeBack, and ComposeEnd capture the decomposition of a formula and the combination of the models of its subformulae and thus affect multiple components.

Decompose. The state of the parent component \(\mathcal {C}\) with formula \(F\) is \(o \) (open). The trail \(I\) neither satisfies nor falsifies \(F\), and \({F}|{}_{I} \) contains no unit clause but can be partitioned into two formulae \(G\) and \(H\) defined over disjoint sets of variables. Subcomponents for \(G\) and \(H\) are created, the number of subcomponents of \(\mathcal {C}\) is set to two and its state is changed to \(f \) (decomposed). Notice that \(\mathcal {C}\) can only be processed further after its subcomponents are closed.

ComposeBack. The state of the component \(\mathcal {C}\) with formula \(F\) is \(f \) (decomposed). Its subcomponents \(\mathcal {C}_G\) and \(\mathcal {C}_H\) with formulae \(G\) and \(H\), respectively, have state \(c \) (closed). Furthermore, \(N \equiv G\) and \(O \equiv H\), hence \({F}|{}_{I} \equiv I \wedge N \wedge O\), which is added to \(M\). This corresponds to enumerating multiple models of \(F\) in one step. This can easily be seen by applying the distributive laws to \(I \wedge N \wedge O\) which gives us a DSOP formula whose disjuncts are satisfying assignments of \({F}|{}_{I} \). The search space has not yet been processed exhaustively (\(\delta (I) > 0\)), backtracking to the second highest decision level occurs, and the state of \(\mathcal {C}\) is changed back to \(o \) (open). Finally, \(\mathcal {C}_G\) and \(\mathcal {C}_H\) are removed from the global state. If \(I\) can not be extended to a model of \(F\), we have \(N = \bot \) or \(O = \bot \), and \(I \wedge N \wedge O = \bot \). Otherwise, \(I \wedge N \wedge O \ne \bot \). Both cases are captured by rule ComposeBack.

ComposeEnd. The state of the parent component \(\mathcal {C}\) with formula \(F\) is \(f \) (decomposed). Its subcomponents \(\mathcal {C}_G\) and \(\mathcal {C}_H\) with formulae \(G\) and \(H\), respectively, are closed. Furthermore, \(N \equiv G\) and \(O \equiv H\), hence \({F}|{}_{I} \equiv I \wedge N \wedge O\), which is added to \(M\). The search space has been processed exhaustively (\(\textsf {decs}{(I)} = \emptyset \)), and the state of \(\mathcal {C}\) is set to \(c \) (closed). Finally, \(\mathcal {C}_G\) and \(\mathcal {C}_H\) are removed from the global state. As in rule ComposeBack, either \(I \wedge N \wedge O = \bot \) or \(I \wedge N \wedge O \ne \bot \).

Example 2

Reconsider Example 1 with variables and defined over \(V\). The execution trace of ACD is shown in Fig. 3. Unaffected components are depicted in gray, and model enumeration by means of chronological CDCL is shown only once in full detail. The execution starts with the root component \(\mathcal {C}_F\) containing \(F\). In step (1), the unit literal \(a\) is propagated, upon which \({F}|{}_{a} \) is decomposed into \((g \vee h)\) and \(G\) creating components \(\mathcal {C}_{(g \vee h)}\) and \(\mathcal {C}_G\) shown in (2). Steps (3) to (6) capture model enumeration by chronological CDCL of \((g \vee h)\), i.e., the computation of a DSOP representation of \((g \vee h)\), after which \(\mathcal {C}_{(g \vee h)}\) is closed. Next, the formula \(G\) is processed by deciding \(b\) in step (7), decomposing \({G}|{}_{b} \) into \((c \vee d)\) and \((e \vee f)\) and creating components \(\mathcal {C}_{(c \vee d)}\) and \(\mathcal {C}_{(e \vee f)}\), respectively, in step (8). The processing of \(\mathcal {C}_{(c \vee d)}\) and \(\mathcal {C}_{(e \vee f)}\) occurs analogously to steps (3) to (6) resulting in the state shown in (9). The results are conjoined with \(b\), which is the trail of \(\mathcal {C}_G\) and under which \({G}|{}_{b} \) was decomposed. Since \(b\) is a decision, it is flipped in (10) to explore its right branch \(\lnot {b}\). The formula \({G}|{}_{\lnot {b}} \) is decomposed into \((\lnot {c} \vee e)\) and \((d \vee f)\) and components \(\mathcal {C}_{(\lnot {c} \vee e)}\) and \(\mathcal {C}_{(d \vee f)}\) are created, as in (11). Their processing, which is not shown, results in the state depicted in (12), and the results are conjoined with the trail of \(\mathcal {C}_G\). Since its trail contains no decision, \(\mathcal {C}_G\) is closed, see (13). The global state now contains the root component and its two subcomponents, which are closed, hence the rule ComposeEnd is executed, and the computation terminates with the closed root component and \(M = a \wedge (g \vee \lnot {g} \wedge h) \wedge (b \wedge (c \vee \lnot {c} \wedge d) \wedge (e \vee \lnot {e} \wedge f) \vee \lnot {b} \wedge (c \wedge e \vee \lnot {c}) \wedge (d \vee \lnot {d} \wedge f)\), where \(M \equiv F\), and which is shown in (14).

Fig. 3.
figure 3

Execution trace of ACD for Example 1.

5 Proofs

For proving correctness, we first show that our calculus is sound by identifying invariants which need to hold in a sound global state and show that they still hold after the execution of any rule. Then we prove that for any closed component it holds that \(M \equiv F\) and that ACD can not get stuck and terminates in a correct state. Showing termination concludes our proof.

Definition 1 (Sound Global State)

A global state \(\mathcal {S}\) is sound if for all its components \(\mathcal {C} = (F,\,V,\,d,\,e,\,I,\,M,\,\delta )^s\) the following invariants hold:

(1):

\(\forall k, \ell \in \textsf {decs}{(I)} \mathrel {.}\tau ({I}, {k})< \tau ({I}, {\ell }) \implies \delta (k) < \delta (\ell )\)

(2):

\(\delta (\textsf {decs}{(I)}) = \{1\}{\ldots }{\delta (I)}\)

(3):

\(\forall n \in \mathbb {N}\mathrel {.}F\wedge \lnot {M}{{\textsf {decs}}_{\leqslant n}(I)} \models {I}_{\leqslant n}\), provided \(\mathcal {C}\) is open or decomposed

(4):

\(M\vee O(I)\) is a d-DNNF, provided \(\mathcal {C}\) is open or decomposed

(5):

\(M\vee F\wedge O(I) \equiv F\)

(6):

\(e > 0\) iff (A) \(e = 2\), (B) \(\mathcal {C}\) is decomposed, (C) \(\mathcal {S}\) contains components \(\mathcal {C}_G = (G,\,\textsf {var}(G),\,d{1},\,e_g,\,J_G,\,N,\,\delta _G)^s\), \(\mathcal {C}_H = (H,\,\textsf {var}(H),\,d{2},\,e_H,\,J_H,\,O,\,\delta _H)^s\), such that \({F}|{}_{I} = G \wedge H\) and \(\textsf {var}(G) \cap \textsf {var}(H) = \emptyset \)

(7):

If \(e = 2\) and \(\mathcal {S}\) contains components \(\mathcal {C}_G = (G,\,\textsf {var}(G),\,d{1},\,0,\,J_G,\,N,\,\delta _G)^c\) and \(\mathcal {C}_H = (H,\,\textsf {var}(H),\,d{2},\,0,\,J_H,\,O,\,\delta _H)^c\), then \({F}|{}_{I} \equiv I \wedge N \wedge O\)

(8):

If \(\mathcal {C}\) is closed, then \(\textsf {decs}{(I)} = \emptyset \)

Invariants (1) - (5) correspond to the ones in our previous work [34]. They say that decisions are ordered in ascending order with respect to their decision level and that every decision level contains a decision literal. They further ensure that literals propagated after backtracking upon finding a model are indeed implied, that no model is enumerated multiple times and that all models are found. Invariant (3) is only useful for open or decomposed components, since \(I\) remains unaltered when a component is closed. Invariant (4) only holds for closed components if \(I(F) = \bot \). Invariants (6) and (7) are concerned with the properties of a parent component and its subcomponents (for the case \(c = 2\)), such as the definition of the component level. Since, given a trail \(I\), \({F}|{}_{I} \) is decomposed into formulae \(G\) and \(H\), we also have that \({F}|{}_{I} \equiv N \wedge O\), where \(N \equiv G\) and \(O \equiv H\). Finally, Inv. (8) says that the trail of a closed component contains no decision.

Lemma 1 (Soundness of the Initial Global State)

The initial global state is sound.

Proof

Due to \(I = \varepsilon \) and \(e = 0\) and since the (root) component is open, all invariants in Definition 1 are trivially met.

Theorem 1 (Soundness of ACD Rules)

The rules of ACD preserve soundness, i.e., they transform a sound global state into another sound global state.

Proof

The proof is carried out by induction over the rule applications. We assume that prior to the application of a rule the invariants in Definition 1 are met and show that they also hold in the target state. The (parent) component in the original state is denoted by \(\mathcal {C} = (F,\,V,\,d,\,e,\,I,\,M,\,\delta )^s\) and in the target state by \(\mathcal {C'} = (F,\,V,\,d',\,e',\,I',\,M',\,\delta ')^{s'}\). Its subcomponents, if there are any, are written \(\mathcal {C}_G = (G,\,\textsf {var}(G),\,d{1},\,e_G,\,J,\,N,\,\delta _G)^s\), \(\mathcal {C}_H = (H,\,\textsf {var}(H),\,d{2},\,e_H,\,K,\,O,\,\delta _H)^s\). Unit, Decide, BackTrue, and BackFalse: Apart from the additional elements \(V\), \(d\), \(e\) and the component state \(s\), the rules are defined as in the former calculus [34]. The arguments given in the proof there apply here as well, and after applying rules Unit, Decide, BackTrue, or BackFalse, Inv.  (1) - (5) hold. Notice that in the proof of Inv. (4), it suffices to replace “DSOP” by “d-DNNF”, since the relevant property here is determinism. Since \(e' = 0\), Inv. (6) and (7) do not apply. An open state is mapped to an open state, hence Inv. (8) holds.

CompTrue and CompFalse: Invariants (1) and (2) hold, since \(I\) remains unaffected. Since \(\mathcal {C'}\) is closed, Inv. (3) and (4) are met. The proof that Inv.  (5) holds is carried out similarly to the proof of Proposition 1 in our previous work [34] for rules EndTrue and EndFalse, respectively. Since \(e' = 0\) and \(I' = I\), Inv.  (6) - (8) hold.

Decompose: The parent component \(\mathcal {C}\) remains unaltered except for \(e' = 2\) and for its state, which becomes \(f \). Both its subcomponents \(\mathcal {C}_G\) and \(\mathcal {C}_H\) are open, and we have \(J_G = J_H = \varepsilon \) and \(e_G = e_H = 0\). Therefore, Inv. (1) - (5) hold. Invariant (6) is satisfied by the definition of rule Decompose. Since \(\mathcal {C'}\) is decomposed and \(\mathcal {C}_G\) and \(\mathcal {C}_H\) are open by definition, Inv. (7) and (8) hold as well.

ComposeBack: It suffices to show that the validity of the invariants for \(\mathcal {C'}\) is preserved, since \(\mathcal {C}_G \) and \(\mathcal {C}_H\) do not occur in the target state. The most recent decision literal is flipped, similar to rule BackTrue. The same argument to the one given there applies, and Inv. (1) and (2) are satisfied. We need to show that \(F \wedge \lnot {(M \vee (I \wedge N \wedge O))} \wedge {\textsf {decs}}_{\leqslant n}( P{K}{\ell }) \models {( P{K}{\ell })}_{\leqslant n}\) holds for all \(n\). The decision levels of the literals in \( P{K}\) do not change, except for the one of \(\ell \), which is decremented from \(e+1\) to \(e\). The literal \(\ell \) also stops from being a decision literal. Since \(\delta ( P{K}{\ell }) = e\), we can assume \(n \le e\). Furthermore, \(F \wedge \lnot {(M \vee (I \wedge N \wedge O))} \wedge {\textsf {decs}}_{\leqslant n}( P{K}{\ell })) \equiv (\lnot {I} \wedge (F \wedge \lnot {M} \wedge {\textsf {decs}}_{\leqslant n}(I))) \vee (F \wedge \lnot {M} \wedge \lnot {(N \wedge O)} \wedge {\textsf {decs}}_{\leqslant n}(I))\), since \(\ell \) is not a decision literal in \( P{K}{\ell }\) and \({I}_{\leqslant e} = P{K}\) and thus \({I}_{\leqslant n} = {( P{K})}_{\leqslant n}\) by definition. By applying the induction hypothesis, we get \(\lnot {I} \wedge F \wedge \lnot {M} \wedge {\textsf {decs}}_{\leqslant n}( P{K}{\ell }) \models {( P{K})}_{\leqslant n}\), and hence \(F \wedge \lnot {(M \vee (I \wedge N \wedge O))} \wedge {\textsf {decs}}_{\leqslant n}( P{K}{\ell }) \models {( P{K})}_{\leqslant n}\). We still need to show that \(F \wedge \lnot {(M \vee (I \wedge N \wedge O))} \wedge {\textsf {decs}}_{\leqslant e}( P{K}{\ell }) \models \ell \), as \(\delta (\ell ) = e\) in \( P{K}{\ell }\) after applying ComposeBack and thus \(\ell \) disappears from the proof obligation for \(n < e\). Notice that \(F \wedge \lnot {D} \models I\) using again the induction hypothesis for \(n = e+1\). This gives us \(F \wedge \lnot {{\textsf {decs}}_{\leqslant e}( P{K})} \wedge \lnot {\ell } \models I\) and thus \(F \wedge \lnot {{\textsf {decs}}_{\leqslant e}( P{K})} \wedge \lnot {I} \models \ell \) by conditional contraposition, and Inv. (3) holds.

For proving that Inv. (4) holds, we consider two cases: (A) \(I \wedge N \wedge O \ne \bot \), i.e., there exists an extension of \(I\) which satisfies \(F\), and (B) \(I \wedge N \wedge O =\bot \), i.e., all extensions of \(I\) falsify \(F\). For both cases, we know that \(I \vee O(I)\) is a d-DNNF.

(A) We need to show that \(M \vee (I \wedge N \wedge O) \vee O( P{K}{\ell })\) is a d-DNNF. Due to \(\delta (I) = e+1\), we have \(O(I) = I \vee R_{\leqslant e+1}(I) = I \vee R_{\leqslant e}(I) \vee R_{=e+1}(I)\). The pending search space of \( P{K}{\ell }\) is given by \(O( P{K}{\ell }) = P{K}{\ell } \vee R_{\leqslant e}( P{K}{\ell })\). But \( P{K} = {I}_{\leqslant e}\) and \( P{K}{\ell } = {I}_{\leqslant e}{\ell } = R_{=e+1}(I)\), since \(\lnot {\ell } \in \textsf {decs}{(I)} \) and \(\delta (\lnot {\ell }) = e+1\). Furthermore, \(R_{\leqslant e}( P{K}{\ell }) = R_{\leqslant e}( P{K})\), since \(\ell \not \in \textsf {decs}{( P{K}{\ell })} \) and \(\delta (\ell ) = e\), hence \(R_{\leqslant e}( P{K}{\ell }) = R_{\leqslant e}(I)\). We have \(O( P{K}{\ell }) = R_{=e+1}(I) \vee R_{\leqslant e}(I)\), hence \(O( P{K}{\ell }) \vee I = O(I)\) and \((M \vee I) \vee O( P{K}{\ell }) = M \vee O(I)\), which is a DSOP and hence a d-DNNF. Now \(I\), \(N\), and \(O\) are defined over pairwise disjoint sets of variables by construction, i.e., \(I \wedge N \wedge O\) is decomposable, and \(M \vee (I \wedge N \wedge O) \vee O( P{K}{\ell })\) is a d-DNNF.

(B) We need to show that \(M \vee O( P{K}{\ell })\) is a d-DNNF. As just shown, \(O( P{K}{\ell }) \vee I = O(I)\). Now \(M \vee O( P{K}{\ell }) = M \vee R_{\leqslant e+1}(I)\). Recalling that \(R_{\leqslant e+1}(I)\) is equal to \(O(I)\) without \(I\) and \(M \vee O(I)\) is a d-DNNF by the premise, \(M \vee O( P{K}{\ell })\) is a d-DNNF as well. Therefore, Inv. (4) holds.

For the proof of the validity of Inv. (5), given \(M \vee F \wedge O(I) \equiv F\), the same two cases are relevant: (A) \(I \wedge N \wedge O \ne \bot \) and (B) \(I \wedge N \wedge O =\bot \).

(A) We have to show that \(M \vee (I \wedge N \wedge O) \vee (F \wedge O( P{K}{\ell })) \equiv F\). From \(O( P{K}{\ell }) \vee I = O(I)\) we get \(M \vee (F \wedge O(I)) = M \vee (F \wedge (O( P{K}{\ell })) \vee I) = M \vee (F \wedge O( P{K}{\ell })) \vee (F \wedge I) \equiv F\). But \(F \wedge I \equiv I \wedge N \wedge O\). Therefore \(M \vee (F \wedge O(I)) \equiv M \vee (F \wedge O( P{K}{\ell })) \vee (I \wedge N \wedge O) = M \vee (I \wedge N \wedge O) \vee (F \wedge O( P{K}{\ell })) \equiv F\).

(B) We must show that \(M \vee (F \wedge O( P{K}{\ell })) \equiv F\). Similarly to (A) we have \(M \vee (F \wedge O(I)) \equiv M \vee (F \wedge O( P{K}{\ell })) \vee (F \wedge I) \equiv M \vee (F \wedge O( P{K}{\ell })) \equiv F\), due to \(F \wedge I \equiv F\). Therefore, Inv. (5) holds after applying rule ComposeBack. We have \(e' = 0\), and \(\mathcal {C'}\) is open, hence Inv. (6) - (8) trivially hold.

ComposeEnd: It suffices to show that after applying rule ComposeBack the invariants are met by \(\mathcal {C'}\), since its subcomponent states \(\mathcal {C}_G \) and \(\mathcal {C}_H\) do not occur in the target state anymore. Due to \(I' = I\) and \(\textsf {decs}{(I)} = \emptyset \) and since \(\mathcal {C'}\) is closed, Inv. (1) - (4) trivially hold.

For proving that invariant (5) holds after applying rule ComposeEnd, i.e., that \(M \vee (I \wedge N \wedge O) \vee (F \wedge O(I)) \equiv F\), the same two cases need to be distinguished: (A) \(I \wedge N \wedge O \ne \bot \) and (B) \(I \wedge N \wedge O = \bot \).

(A) From \(\textsf {decs}{(I)} = \emptyset \), we get \(O(I) = I\) and \(F \wedge O(I) = F \wedge I\). Recalling that \(F \wedge I \equiv I \wedge N \wedge O\), we obtain \(M \vee (I \wedge N \wedge O) \vee (F \wedge O(I)) \equiv M \vee (F \wedge O(I)) \equiv F\) by the premise.

(B) We have \(M \vee (I \wedge N \wedge O) \vee (F \wedge O(I)) = M \vee (F \wedge O(I)) \equiv F\) by the premise, and Inv. (5) holds after executing rule ComposeEnd. Invariants (6) - (8) trivially hold, due to \(e' = 0\) and \(I' = I\) and hence \(\textsf {decs}{(I')} = \emptyset \).

Corollary 1 (Soundness of ACD Run)

ACD starting with an initial global state is sound.

Proof

The initial state is sound by Lemma 1, and all rule applications lead to a sound state according to Theorem 1.

Lemma 2 (Correctness of Closed Component State)

For any closed component \((F,\,V,\,d,\,0,\,I,\,M,\,\delta )^c\) it holds that \(M \equiv F\).

Proof

Follows from Theorem 1, proof of Inv. (5) for rules CompTrue, CompFalse, and ComposeEnd, which are the only rules closing a component.

Theorem 2 (Correctness of Final Global State)

In the final global state of ACD, \(M \equiv F\) holds.

Proof

Correctness of the closed root component follows from Lemma 2. We need to show that the final global state contains exactly the closed root component. The initial global state consists of the open root component. Additional components are created exclusively by rule Decompose, and a parent component state can only be closed by rule ComposeEnd, which also removes its subcomponents from the global state. Hence the root component can only be closed if it has no subcomponents. But since the initial global state contains exclusively the root component, the final global state contains only the closed root component.

Theorem 3 (Progress)

ACD always makes progress.

Proof

The proof is conducted by induction over the rules. We show that as long as the root component is not closed, a rule is applicable. For the case , where \(\mathcal {C} = (F,\,V,\,d,\,0,\,I,\,M,\,\delta )^o\) has no subcomponents, the proof is identical to the one showing progress in our previous work [34] replacing EndTrue with CompTrue and EndFalse with CompFalse, and by checking whether the preconditions for rule Decompose are met if rule Unit is not applicable and before taking a decision. Now let the global state be given by where \(\mathcal {C} = (F,\,V,\,d,\,2,\,I,\,M,\,\delta )^f\) is decomposed. Due to Inv. \(\textsf {(6)} \), \(\mathcal {S}\) contains \(\mathcal {C}_G = (G,\,\textsf {var}(G),\,d{1},\,e_G,\,J_G,\,N,\,\delta _G)^s\) and \(\mathcal {C}_H = (H,\,\textsf {var}(H),\,d{2},\,e_H,\,J_H,\,O,\,\delta _H)^s\) such that \({F}|{}_{I} = G \wedge H\) and \(\textsf {var}(G) \cap \textsf {var}(H) = \emptyset \). Assume \(s = c \) for both \(\mathcal {C}_G\) and \(\mathcal {C}_H\) . If \(\textsf {decs}{(I)} = \emptyset \), rule ComposeEnd is applicable. Otherwise, similarly to rule BackTrue, we can show that all preconditions of rule ComposeBack are met. If instead for at least one of \(\mathcal {C}_G\) and \(\mathcal {C}_H\), the non-closed component(s) are processed further, and as soon as both \(\mathcal {C}_G\) and \(\mathcal {C}_H\) are closed, rule ComposeEnd or ComposeBack can be applied. This proves that ACD always makes progress.

Theorem 4 (Termination)

ACD always terminates.

Fig. 4.
figure 4

Rule applications lead to smaller global states.

Proof

We need to show that no infinite sequence of rule applications can happen. To this end, we define a strict, well-founded ordering \(\succ _{\textsc {ACD}} \) on the global states and show that \(\mathcal {S} \leadsto _{\textsf {R}} \mathcal {T}\) implies \(\mathcal {S} \succ _{\textsc {ACD}} \mathcal {T}\) for all \(\mathcal {S}, \mathcal {T} \in S\) and rules \(\textsf {R}\) in ACD. Global states are sets of components, and \(\succ _{\textsc {ACD}} \) is the multiset extension of a component ordering \(\succ _{\textsf {c}} = (\succ _{\textsf {cl}}, \succ _{\textsf {tr}}, \succ _{\textsf {cs}} \)), where \(\succ _{\textsf {cl}} \), \(\succ _{\textsf {tr}} \), and \(\succ _{\textsf {cs}} \) are orderings on component levels, trails, and component states, respectively. We want to compare trails defined over the same set of variables \(V\), and to this end we represent them as lists over . A trail \(I = \ell _1{\ldots }{\ell _k}\) defined over \(V\), where \(k \le |{V}|\), is represented as \( [l_1]{\ldots }{l_k}{2}{\ldots }{2}\), where \(l_i = 0\) if \(\ell _i\) is a propagation literal and \(l_i = 1\) if \(\ell _i\) is a decision literal. The last \(|{V}|-m\) positions with value \(2\) represent the unassigned variables. Trails defined over the same variable set are encoded into lists of the same length. This representation induces a lexicographic order \(>_{\textsf {lex}} \) on trails, and we define \(\succ _{\textsf {tr}} \) as the restriction of \(>_{\textsf {lex}} \) to , i.e., we have \(t_1 \succ _{\textsf {tr}} t_2\) if \(t_1 >_{\textsf {lex}} t_2\). The ordering \(\succ _{\textsf {tr}} \) is well-founded, its minimal element is \( [0]{\ldots }{0}\). The component state takes values in , and we define \(\succ _{\textsf {cs}} \) as \(>_{\textsf {lex}} \), i.e., \(s_1 \succ _{\textsf {cs}} s_2\) if \(s_1 >_{\textsf {lex}} s_2\). The minimal element of \(\succ _{\textsf {cs}} \) is \(c \), hence \(\succ _{\textsf {cs}} \) is well-founded. Given two component levels \(d_1\) and \(d_2\), we define \(d_1 \succ _{\textsf {cl}} d_2\) if \(\textsf {length}{(d_1)} < \textsf {length}{(d_2)} \). This may seem counterintuitive but is needed to ensure that the execution of rule Decompose results in a smaller state, since both the component state and the trail of the new subcomponents are of higher order than those of their parent. To see that \(\succ _{\textsf {cs}} \) is well-founded, recall that we consider finite variable sets. Their size provides an upper limit on the length of the component level representation and a minimal element of \(\succ _{\textsf {cs}} \).

Now we define the component ordering \(\succ _{\textsf {c}} = (\succ _{\textsf {cl}}, \succ _{\textsf {tr}}, \succ _{\textsf {cs}})\). Let two components be \(\mathcal {C}_1 = (d_1){t_1}{s_1}\) and \(\mathcal {C}_2 = (d_2){t_2}{s_2}\). We have \(\mathcal {C}_1 \succ _{\textsf {c}} \mathcal {C}_2\) if \(\mathcal {C}_1 \ne \mathcal {C}_2\) and \(d_1 \succ _{\textsf {cl}} d_2\) or \(d_1 = d_2\) and either \(t_1 \succ _{\textsf {tr}} t_2\) or \(t_1 = t_2\) and \(s_1 \succ _{\textsf {cs}} s_2\). Clearly \(\succ _{\textsf {c}} \) is well-founded, since \(\succ _{\textsf {tr}} \), \(\succ _{\textsf {cs}} \), and \(\succ _{\textsf {cl}} \) are well-founded. For two global states \(\mathcal {S}\) and \(\mathcal {T}\), we have \(\mathcal {S} \succ _{\textsc {ACD}} \mathcal {T}\) if \(\mathcal {S} \ne \mathcal {T}\) and for each component \(\mathcal {C}\) such that \(\mathcal {C}\) is larger in \(\mathcal {T}\) than in \(\mathcal {S}\) with respect to \(\succ _{\textsf {c}} \), \(\mathcal {S}\) contains a component \(\mathcal {C}'\) that is larger in \(\mathcal {S}\) than in \(\mathcal {T}\). Since \(\succ _{\textsf {c}} \) is well-founded, also \(\succ _{\textsc {ACD}} \) is well-founded. Figure 4 shows that each rule application leads to a smaller global state, concluding our proof.

6 Generalization

The generalized rules are listed in Fig. 5. In our generalized framework, we have \({F}|{}_{I} = \bigwedge _{i=1}^{n}G_i\), and \(\textsf {var}(G_i) \cap \textsf {var}(G_j) = \emptyset \) for and \(i \ne j\) (rule DecomposeG). Similarly to their equivalents in ACD, rules ComposeBackG and ComposeEndG are applicable if all subcomponents are closed.

Fig. 5.
figure 5

Generalized transition rules.

7 Discussion

We have presented Abstract CNF2dDNNF, or ACD for short, a formal framework for compiling a formula in CNF into d-DNNF combining CDCL-based model enumeration with chronological backtracking [34] and dynamic component analysis [4]. Conflict-driven clause learning enables our framework to escape regions without solution early, and chronological backtracking prevents multiple model enumeration without the need for remembering already found models using blocking clauses, which slow down unit propagation. However, the absence of blocking clauses also prevents the use of restarts. If exclusively the rules Unit, Decide, BackTrue, BackFalse, CompTrue, and CompFalse are used, a DSOP representation of \(F\) is computed. Unit propagation is prioritized due to its potential to reduce the number of decisions and thus of right branches to be explored. Favoring decompositions over decisions may also shrink a larger part of the search space. Our framework lays the theoretical foundation for practical All-SAT and \(\#\)SAT solving based on chronological CDCL. Any implementation which can be modeled by ACD exhibits its properties, in particular its correctness, which has been established in a formal proof.

Comparison with Available Tools. There exist other knowledge compilers addressing d-DNNFs. We want to mention c2d [20], Dsharp [37], and D4 [30], which also execute an exhaustive search and conflict analysis. However, our approach differs conceptually from these tools in several ways. The most prominent ones are the use of CDCL with chronological backtracking [33, 38] instead of CDCL with non-chronological backtracking and the way the d-DNNF is created. Our method generates DSOP representations of formulae which can not be decomposed further by an exhaustive (partial) model enumeration and then combines the result, while the tools mentioned above generate the d-DNNF by recording the execution trace as a graph [26, 27]. As ACD, both D4 and Dsharp adopt a dynamic decomposition strategy, while c2d constructs a decomposition tree which it then uses for for component analysis.

Future Research Directions. We plan to implement a proof of concept of our calculus in order to compare the size of the returned d-DNNF with the ones obtained by c2d, D4, and Dsharp. For dynamic component analysis, one could follow the algorithm implemented in COMPSAT [6], while dual reasoning [32] and logical entailment [35] enable the detection of short partial models. This is particularly interesting in tasks where the length of the d-DNNF is crucial. Dual reasoning has shown to be almost competitive on CNFs if the search space is small, we therefore expect that component analysis boosts its performance. The major challenge posed by the second approach lies in an efficient implementation of the oracle calls required by the entailment checks. It would be interesting to investigate the impact of dynamic component analysis on a recent implementation [46] of model enumeration by chronological CDCL [34]. Cache structures, being an inherent part of modern knowledge compilers and \(\#\)SAT solvers [11, 16, 19, 20, 30, 31, 37, 41, 42, 47, 49] due to their positive impact on solver efficiency [1], should be added to any implementation of our framework. Finally, an important research topic is that of optimizing the encoding of a formula making best use of component analysis [14]. Related to this question is whether formulae stemming from practical applications are decomposable in general.