Dialogue systems have been originally proposed for knowledge-based systems [2] and could be used to realize joint decision making. Human interaction can serve as a model. Here, one human might ask another to explain his or her decision. For example, one medical expert might ask another for the reasons behind a pT3 diagnosis (see Fig. 2). The given explanation can be accepted by the other person or not. In case of rejection, it can be indicated which parts of the given explanations are not acceptable. Together the discussion partners can find an alternative explanation. We propose that this dialogue-based, incremental process should be captured by explainable AI methods. Such mutual explanations are cooperative, interactive, and incremental acts of information exchange between humans and machines with the goal to improve the joint performance of the involved partners in classification problems. We propose that the process of explanation refers to providing arguments [20] that make simple and complex relations, which apply to the domain of interest, explicit. It further refers to integrating corrective explanations into existing internal models in order to adapt these [10].
A model of such a mutual explanation system is given in Fig. 1: Starting with an initial ILP model, a new example e is classified. The class decision is presented to the human who can confirm it or ask for an explanation. The explanation can be accepted or not. In case of rejection, the human can correct the explanation. The correction together with the new class label are integrated to adapt the model. While it is possible that a correct classification can be associated with a wrong or insufficient explanation, we focus on correcting explanations associated with erroneous class decisions. The proposed approach is applicable to both cases.
A wrong class decision can be either a miss or a false alarm. In machine learning, this can be attributed to overly specific or overly general model [6, 19]. In ILP, a learned model \(\mathcal {M}\) for a single target predicate consists of first order rules R of the following form:
$$\begin{aligned} H \leftarrow l_1 , \ldots , l_m. \end{aligned}$$
where the head of the rule is an atom and the body is a conjunction of literals. Rules are defined over variables. An instance is a conjunction K of ground literals. An instance is classified as member of the target concept, if there exists a substitution \(\theta \) of variables in a rule \(R \in \mathcal {M}\) such that \(K \subseteq body (R)\theta \).
As described in De Raedt [5] theta-subsumption can be considered to be a constraint satisfaction problem. Hence, human corrections of explanations can involve adding or deleting literals or restricting or enlarging the scope of arguments.
More formally, given a conjunction of literals (\(l_{1}\)\(\wedge \) ...\(\wedge \)\(l_{n}\)) from \( body (R)\) and a conjunction of boolean constraints (\(c_{1}\)\(\wedge \) ...\(\wedge \)\(c_{n}\)), a substitution \(\theta \in \{\top , \bot \}\) needs to be found for every literal, such that the resulting boolean formula \(L_{i} \wedge C_{n}\) evaluates to \(\top \), if \(L_{i}\) is a valid clause to be added to the theory, \(\bot \) otherwise. Each \( body (R)\) that satisfies \(C_{n}\) can be added to the theory, if it has the best score with respect to Aleph’s evaluation setting. Given a literal’s set of arguments (\(a_{1} \wedge \ldots \wedge a_{n}\)), a domain \(D(a_{i})\) for every \(a_{i}\) and a conjunction of numerical or set constraints, a substitution \(\theta \) needs to be found, such that \(D(a_{i})\)\(\ge _{g}\)\(D(a_{i} \theta )\) for overly general clauses and \(D(a_{i} \theta )\)\(\ge _{g}\)\(D(a_{i})\) for overly specific clauses. For overly specific clauses, \(\theta \) substitutes constants with variables. For overly general clauses, \(\theta \) substitutes variables with constants or different variables, which are already present in the set of arguments.
An overly general model can result in a false alarm, erroneously classifying an instance as member of the target concept—such as tumor class pT3 or pain. A human expert might introduce an additional literal or restrict the scope of a predicate. For instance, a rule
is too general and can be restricted by introducing
A rule might specify the size of a tumor in millimeters
which also can be restricted in case of over-generality of a rule by requiring the value to be larger than 5.
An overly specific model can result in misses. For instance, the rule
excludes tumors from class pT3 where tumor and muscle tissue are not disjoint. Likewise, specific values, for instance, the size of a tumor, might be too restrictive. Decreasing the minimum size of a tumor makes a rule more general.
In Aleph, user-defined constraints can be applied to guide the generation of alternative clauses [30]. To make the interaction with our system easier for medical experts, they can mark parts of an explanation, which are then transformed into proper constraint syntax. For example, if we require a clause to contain some predicate p with arguments X and a, where X is a variable and a is a constant, a typical constraint is represented as follows:
The head of a constraint is set to false. This way, all clauses evaluate to false, where the goals in the body of the constraint are satisfied. The constraint above expresses that a body containing p must occur in a clause. The set of user-defined constraints and the current clause are combined into a boolean formula for SAT solving as well as unification is performed. Aleph then generates new rule candidates, considering only the ones which satisfy the constraints for theory construction.
We conducted a first experiment to evaluate our mutual explanation approach (see Finzel [10] for details). We generated a small artificial data set for the colon cancer domain and introduced erroneous class labels which resulted in false positives. We iteratively applied boolean constraints for corrections of erroneous explanations at the clause level. At the literal level we applied set and numerical constraints. All constraints were generated from user feedback via an explanation interface as shown in Fig. 2.
Applying the constraints led to a specialization of the induced theory and thus the exclusion of false positives. Results further indicate that introducing constraints can help to decrease the necessary number of corrections. However, corrections can result in higher computational effort during search. This preliminary evaluation can be seen as a first proof of concept. We currently are conducting an evaluation with a larger data set where we assess the reduction of errors and computational time in a systematic way.