1 Introduction

Up to now, ML systems are mostly used in areas, where wrong decisions only have minor consequences. The increasing performance of ML compared to classical systems especially for perceptions tasks in natural complex environments gives hope for, e.g., driver-less vehicle operation in foreseeable future. The hopes of using ML also for perception tasks in areas of higher risk currently may not be fulfilled, due to the required trustworthiness of the systems including assured high quality of the used data sets may not be proven. To give a legal framework for the application of ML systems, the European AI act [6] is currently under development. Simultaneously, there are multiple projects from research and industry working on the topic of ML Systems in high-risk areas like KI-Absicherung [3, 16] and safe.trAIn [26]. High-risk ML systems have to fulfill the requirements according to [6] Chapter 2 "REQUIREMENTS FOR HIGH-RISK AI SYSTEMS" Article 10 "Data and data governance" Point 3:

Training, validation, and testing data sets shall be relevant, representative, free of errors and complete.

In this paper, we present a novel approach to visually represent and quantify data quality, and demonstrate this approach with several different application examples. The aim is to provide new tools to understand and manage the data quality. The approach is a part of the QUEEN-method (Qualitätsgesicherte effiziente Entwicklung vorwärtsgerichteter künstlicher Neuronaler Netze, quality-assured efficient development of neural networks) [30] a high-level procedure for quality-assured development of neural networks. QUEEN gives guidance for developing explainable artificial intelligence (XAI) based on the development of dedicated prepossessing. It is supported with quality indicators that are used as key performance indicators for the achieved complexity reduction for the given task. Quality assuring the processing of every single step before the data are used as input for a neural network supports the trustworthiness for the data bases and is combined with lightweight AI systems that can be interpreted in detail. QUEEN defines two data quality assurance methods called QI\(^2\) (integrated quality indicator) [8] and ECS (equivalent classes sets) [24]. In this paper, we want to show the mathematical bases as well as the applications of QI\(^2\) and its underlying methods in terms of data quality assurance. The method quantifies the complexity of input–output relationship defined with a data set. The term complexity herein is representative for a measurement of the amount of non-linearities in input–output relationships. It is based on the pairwise measurement of input and output distances what makes it attractive for a wide variety of applications due to simply adapting the distance metric. The ECS is not part of this paper but is covered in the submission [25].

The QI\(^2\) and its respective methods are a handy tool in terms of quality assurance due to a low dimensional representation of data. It quantifies neighborhood input–output relationship behaviors over a set of data points. High-dimensional anomalous structures and relationships in data are represented by anomalous structure in a simple visualization. It is possible to directly and accurately address several quality aspects—e.g., global linearity, outlier, simple and difficult sub-structures, clustering structure, inconsistent data, discontinuities—by simply interacting with its visual representation.

2 Related work

Data quality in general is a widely researched topic due to the growing necessity and usage of data in our every day life. But there are still a lot of problems to be tackled. In [28] Data quality is defined with respect to the intended use of the consumer. Although data should be intrinsically good, they state that data quality needs to be a context-dependent term to be appropriate for the desired task. The term of data quality is furthermore split into many dimensions like accuracy, consistency, completeness, safety, timeliness, etc. in [23]. [9] considers the ISO/IEC 25012, an accepted industry standard for data quality and gives a model based approach. In addition to that, [17] states a separation into subjective and objective quality assessment with three functional forms. The diversity of the definition of data quality has to be resolved in the future to give a guidance to implement the European AI Act. A consensus seems to exist, that data quality has intrinsic properties—accuracy, completeness, consistency, etc—as well as properties that are more dependent on use—timeliness, presentation quality, etc.

As a first glance in data quality assurance, descriptive statistics [11] is used. Herein, statistical methods for measuring central tendencies, dispersion, and association are used for basic understanding of variables in data. The most common displaying method for those statistics is visualization via scatter plots and histograms. Our proposed method extends descriptive statistical methods with a new visualization that gives deeper inside to the spatial relationships in the typically high-dimensional data to support dedicated quality assurance and enables direct interactions between different visualizations of the data.

Dimensional reduction methods like PCA [13], tSNE [14] or UMAP [15] are a further visually representative quality assurance. These methods map the high-dimensional data into low dimensions that may be easily analyzed and understood by humans. Nevertheless, they suffer from a high loss of information and might impose misleading interpretations. The herein proposed method of QI\(^2\) is a human interpretable visual representation of the data according to neighborhood relationships. It, therefore, represents high-dimensional information of the data that are necessary for efficient quality assurance. It does not contain the data itself in the visualization but it is computed on the original values and the structural behavior.

Further metrics claiming approaches for general data quality assurance try to cover as many dimensions of data quality as possible by testing data against predefined rules and assumptions. In contrast to descriptive statistics, those methods do not define data quality due to a statistical and distributional structure that is met by the overall data. They assess data to meet certain rules and assumptions. [12] published the pointblank R package for agent-based data quality assurance. This package contains quality assessments on a higher-level basis with testing values in specific columns or rows against defined functions. One can test if the data are correct in terms of the values either being greater, lower, equal, or between certain values or follow structural or informational rules defined by the user. Another general quality assurance method is DEEQU published in [21] and [22]. This package allows the user to define specific assumption based unit tests for data at large scale. Similar to the previous one, this package also adds row and column based checks to a pipeline for data quality assurance. In addition to unit tests on a single data set, there is the possibility of anomaly detection over time. A further assumption based approach for data quality assurance is shown in [10]. The probability-based method herein outputs a value representing the probability of a data set being free from internal errors with respect to given rules. In comparison to the previous rule-based approaches, this method does not apply rules row or column based. The rules defined for this assurance are more like relationships between individual columns of data sets representing correlations. Despite a more complex application of rule-based checks all mentioned methods are heavily assumption and rule based and, therefore, require high knowledge of the data set and accurate assumptions. Assumptions on data although will never allow full coverage of quality dimensions because one can not know what one does not know. Due to a representative visualization of input–output relationship behavior, our proposed method does not need any assumptions for certain quality aspects. It can be used without any knowledge about the data and does not need thresholds or possible coincidences between data dimensions.

For deeper quality assessment, more specific methods need to be taken into account. Herein, methods developed for a specific purpose will be the key. In the field of outlier detection, for example, there exist a variety of different approaches all tackling the same problem. A first strategy is using density-based outlier detection algorithm [2]. As a first glance, this method outputs a value similar to a probability determining if a data point being an outlier. This value is computed based on how locally isolated the data point is. For this, the local density of each point is compared to the local density of their k-nearest neighbors using the local reachability density (LRD). Another clustering-based approach uses DBSCAN [5] as clustering algorithm and repeats a specific clustering and cluster merging algorithm to a state in which every cluster has its own \(\epsilon\) value regarding DBSCAN. After that, minPts value of DBSCAN will be computed for each cluster with respect to the smallest minPts of that cluster. Regarding minPts, clusters will be classified as anomalous [27].

In further development, density-based clustering was used as a preprocessing in outlier detection combined with inter-cluster distance-based classification of clusters as anomalous. Methodologically similar approaches differing in clustering, distance computation, and anomaly detection are [7, 19, 20]. The first one is using the fixed width clustering algorithm for finding clusters in a data set followed by computing inter-cluster distances. Clusters are classified as anomalous by looking at the average inter-cluster distance and its deviation from the mean inter-cluster distance. The second one above is using the previously mentioned DBSCAN clustering algorithm in combination with a anomalous classification based on inverse distance weighting (IDW). The third one is using OPTICS [1] algorithm for clustering and an inter-cluster distance-based anomalous classification with kriging methods. The methods of QI\(^2\) could as well be seen as a clustering-based analysis method but differs from the mentioned methods due to the fact that outlier detection is not based on thresholds and clustering as preprocessing directly. Clustering as preprocessing can lead to loss in information regarding individual data points. Furthermore the proposed method gives a wider variety in quality assurance than only tackling one specific problem.

3 QI \(^2\): the quality indicator

Since the algorithm presented visually represents the relationships between input and output changes within a data set P, there must first be a division of the data set into input space and output space. Therefore, every data point can be described by

$$\begin{aligned} \vec {p^T} :=\begin{pmatrix} vi_{1} \ldots vi_{I} vo_{1} \ldots vo_{O} \end{pmatrix} \end{aligned}$$
(1)

with I values of the input space \(vi_{i} \epsilon \mathbb {R}\) and O values of the output space \(vo_{o} \epsilon \mathbb {R}\).

With this separation the method of QI\(^2\) can be applied. First of all, it is necessary to calculate the pairwise distances regarding input space and output space throughout every possible pair \(P^2 :=\{x :=(\vec {p_{1}}, \vec {p_{2}}) \mid \vec {p_{1}}, \vec {p_{2}} \epsilon \mathbb {R}^{(I+O)}\}\) of data points. For this every possible metric M can be applied. For example, euclidean or cosine distance for scalar values and SSIM [29] or mutual information [18] for images. Despite using a certain metric for the input space, the output space can have a different metric applied. Take as an example a quality assurance of an image classification data set. In the input space (images), the SSIM can be applied to compare the structural information as distance metric but in the output space (classes as scalar values), the euclidean distance can be applied.

$$\begin{aligned} d_{RI}(x) :=M(x), \end{aligned}$$
(2)
$$\begin{aligned} d_{RO}(x) :=M(x). \end{aligned}$$
(3)

After that the distances will be normalized by mean for comparability. In the following, this will be shown exemplary for the input space.

$$\begin{aligned} d_{NRI}(x) :={d_{RI}(x)}\Big /{\frac{\sum _{y \epsilon P^2} d_{RI}(y)}{|P^2|}}. \end{aligned}$$
(4)

Having the pairwise normalized distances in input- and output space, it is possible to compute a first representative value for a data set described by the data points taken into account. This value describes the complexity and quality of this data set by comparing the relationship between output changes following on certain input changes.

$$\begin{aligned} QI^2R(P) :=\frac{1}{|P^2|} \sum _{x \epsilon P^2} (d_{NRI} - d_{NRO})^2. \end{aligned}$$
(5)

This representative value gives a first shot in quality assurance of a data set by describing the non-linearity. The higher the value, the more non-linear parts are in the data set. A second handy property of this value is that it describes a two-dimensional data set with random normal distribution with a value of one. If the QI\(^2\)R of a data set is higher than one, the input–output behavior of the data set is more complex than a random normal distribution. This occurs, e.g., if parts of the data set are more likely to have more linear relationships between input and output but are separated by a sudden change in direction of the gradient.

The next step in quality assurance of a data set is a local complexity representation. Therefore, the QI\(^2\)R can be repeatedly computed over subsets of the original data set. For this, the data set needs to be sub sampled by, e.g., building subsets based on increasing k-nearest neighborhoods around every data point. This means for every data point in the data set there will be a number of subsets depending on how much neighbors should be taken into account. Another possibility is to divide the data set into subsets with respect to a high-dimensional sphere around each data point with increasing diameter. With those subsets, it is now possible to compute a matrix of local QI\(^2\) (MLQI\(^2\)).

$$\begin{aligned} mlqi^2_{i,k}(P) :=QI^2R(KNN(P, p_{i}, k)), \end{aligned}$$
(6)

\(KNN(P, p_{i}, k)\) represents the buildup of subsets based on k-nearest neighbors around the respective data point \(p_{i}\) in the data set P. The computation of MLQI\(^2\) considers every pairwise distance between possible pairs of points in the subset.

Visualizing a matrix human interpretable is hardly possible in this case. Thus, a visualization via a histogram of local QI\(^2\) (HLQI\(^2\)) was chosen.

$$\begin{aligned} hlqi^2_{v,k}(P) :=\sum _{i=1}^{|P|}{I^3({mlqi^2}_{i,k}(P), v)\cdot {blqi^2}_{i,k}(P)}. \end{aligned}$$
(7)

The term \(blqi^2_{i,k}\) (Boolean matrix of local QI\(^2\), BLQI\(^2\)) is responsible for preventing boundary effects by comparing neighborhoods. It checks whether the QI\(^2\)R for the current neighborhood has already been computed, since this neighborhood already exists on the basis of another data point. If it has already been computed, this calculation will be dropped in the computation of \(hlqi^2_{v,k}\).

$$\begin{aligned} blqi^2_{i,k}(P) :={\left\{ \begin{array}{ll} 1, &{} \text {for }\forall {j<i} \\ &{} | KNN(P,p_{j},k) \ne KNN(P,p_{i},k)\\ 0, &{} \text {else}. \end{array}\right. } \end{aligned}$$
(8)

\(I^3(h,v)\) is responsible for sorting the MLQI\(^2\) into the desired histogram by checking if the receiving value h lies in the bin \([v|v+1)\).

$$\begin{aligned} I^3(h,v) :={\left\{ \begin{array}{ll} 1, &{} \text {for } v \le \frac{h-min_{hi}}{binsize_{hi}} < v+1 \\ 0, &{} \text {else}. \end{array}\right. } \end{aligned}$$
(9)

As a last step, the histogram is scaled to its maximum value in every k so that it contains relative values for the bins and it is gamma calibrated due to visual reasons for lower values (scaled histogram of local QI\(^2\), SHLQI\(^2\)).

$$\begin{aligned} shlqi^2_{v,k}(P) :=\left( \frac{hlqi^2_{i,k}(P)}{\sum _{s=1}^{|P|}{hlqi^2_{s,k}(P)}}\right) ^{gamma_{hi}}. \end{aligned}$$
(10)

3.1 Benefits of QI \(^2\)

The most important advantage of this method is a compressed and representative visualization of local input–output relationship behavior through various different subsets based on every data point. By this, the visualization and knowledge of the computational structure can be used to exactly identify certain interesting data points and quality aspects simply by interaction between the histogram and the data set. Those aspects are, e.g., outliers, linear subtasks, discontinuities, and many more. As stated above in comparison to quality assurance with a dimension reduction algorithm this method gives a more accurate representation due to no possible distortion in data relationships. This is simply because of directly considering the original values for every data point as it is represented in the data set in the SHLQI\(^2\) plot. Nevertheless, a dimension reduction algorithm can be used to visualize the data set human interpretable. In combination with the interaction between the SHLQI\(^2\) and the data visualization, it is possible to get a deep dive into the structure and input–output relation of the data set. Furthermore the SHLQI\(^2\) can be used for quantitative quality assurances due to analysis based on quantitative requirements. An Example of a visualization of SHLQI\(^2\) and a dimensional reduced two-dimensional visual representation of MNIST [4] test set is given in Fig. 1.

3.2 Using QI \(^2\) to understand data quality

The methods of QI\(^2\) can be used to determine certain quality aspects in data sets as stated in the introduction. To assure quality, two separate analyses are possible

  • global analysis and

  • local analysis.

In a first glance of a global analysis, the value of QI\(^2\)R as well as its density function and the cumulative distribution can be used to determine a global complexity. Those methods are calculated based on either the whole data set or based on the sum of increasing clusters around each data point. For further local data quality assurance, the SHLQI\(^2\) can be used. The most important aspect in quality assurance with SHLQI\(^2\) is that the visualization can be used as a human interpretable representation of a high-dimensional data set and, therefore, has to be treated like that. It is necessary to watch out for anomalous and differing areas in the SHLQI\(^2\) from the ’general’ structure of the histogram. Since it analyses the input–output behavior of a data set by representative visualization, it detects anomalous or interesting structures with anomalous visualization. Interesting or anomalous structures in data can be both positive and negative in terms of quality and need a separate treatment afterward. Nevertheless, there are certain characteristics due to the computation of the MLQI\(^2\) that are not straightforward and, therefore, differ in the visualization of SHLQI\(^2\). For this differentiation data sets need to be classified regarding their task, approximation or classification due to huge differences in structure and analyzing methods between these two tasks.

Fig. 1
figure 1

UMAP and SHLQI\(^2\) representation of the MNIST handwritten digits test data set

Fig. 2
figure 2

Comparison of SHLQI\(^2\) of approximation and classification task

3.2.1 Approximation tasks

For this kind of task, the histogram will be build up between bins \(mlqi^2_{i,k} \epsilon [0|2.5]\) in most cases. It has a wider coverage of different bins in the local area and will decrease in coverage toward one single value at maximum k. There are several different characteristics of this histogram that can be interpreted directly. As mentioned above, the final value of the SHLQI\(^2\) at maximum k describes the overall complexity of the whole data set. This is a good starting point in terms of quality assessment due to the fact that for every dimensionality of the input data space there is a specific value for a randomly normal distributed data set. As mentioned above for one dimension in input space, it is the value one. For increasing dimension, the value decreases due to the curse of dimensionality. If the final value of the SHLQI\(^2\) at maximum k is higher than this, the overall data set has a more complex input–output relationship than a randomly normal distribution. In the case of Fig. 2a, this is the case due to the sudden change in direction at \(x=200\). A second handy feature of the SHLQI\(^2\) is the higher and lower values in local areas. If there are many representative chunks in higher bins in local areas, the data set has locally complex areas that need to be treated externally. In contrast to that, many chunks in lower bins means many local areas that have a more or less linear relationship between input and output. Holes in the histogram represent highly complex areas in the data set like sparsely covered input space, sudden changes in data, outlier, and so on.

3.2.2 Classification tasks

For classification tasks, the structure of the SHLQI\(^2\) is completely different from the one for approximation tasks. First of all, its general structure is similar throughout different data sets. By this, one can directly differentiate between a classification and approximation task just by looking on the histogram. The structure of the SHLQI\(^2\) for classification tasks follows strict rules due to the computation based on characteristics of those tasks. The most specific characteristic is the steep rises of complexity toward values way higher than 2.5 in combination with sudden drops of those rises at a data set specific location. Those steep rises and sudden drops are caused by the normalized distance computation (Eq. 4) in the output space. In terms of a classification, the distances between same classes is zero due to the definition of distance metrics. If the computation gets to a point, where a data point with a different class is the next neighbor within the subset, there is a sudden change in output distances. According to the computational structure considering every pairwise distance, there are now a few distances \(d_{RI}\ne 0\). Lets take a representative example. Let the subset P around \(p_i\) be a set of 100 data points. within this set there are 99 points with the same class as \(p_i\) and one point with a different class. Therefore, this data point with a different class has a pairwise distance to every other point differing from zero, e.g., \(d_{RO}(x)=1\). Now, \(|P^2|=100 \cdot 100=10.000\) possible data point pairs. Every pairwise distance \(d_{RO}\) will be divided by \(\frac{\sum _{y\epsilon |p^2|}{d_{RO}(y)}}{|P^2|} = \frac{100}{10.000} = 0.01\) according to Eq. 4. In comparison to the values of \(d_{NRI}\), the result of Eq. 5 will be a high value due to \(d_{NRO}(y)>> d_{NRI}(y)\) for all pairs y for which the two data points have two different classes. The sudden drop in values of \(\hbox {SHLQI}^{2}\) is caused by a further data point with a different class from the major class interfering with the subset by incrementing the neighborhood. By this, the denominator of Eq. 4 is now nearly doubled due to double the amount of pairwise distances \(d_{RO} \ne 0\) compared to a slight rise in the amount of possible data point pairs.

Another characteristic of the SHLQI \(^2\) for classification tasks is a specific area for local k within the bins \(mlqi^2_{i,k} \epsilon (1|2)\). Herein, most complexities are located in a \(e^{-x}\)-like function representing homogeneous groups of classes. The smaller the value of this noticeable are, the denser the clusters. By adding a weak numerical stabilizer in the denominator in Eq. 4 in order to make it numerical stable in terms of comparing identical values the input–output relationship of homogeneous cluster behave somewhat similar to a random distribution. The behavior can be described as randomly distributed inputs referring to one identical output. The darker this area, the more subsets are local homogenous clusters representing a single class.

4 Exemplary data quality assurance

As stated before, the SHLQI \(^2\) can be used to determine certain quality aspects in data sets with respect to a local analysis. Certain aspects need special treatments regarding the identification via interaction with SHLQI \(^2\). We want to present some exemplary identifications with exemplary and state-of-the-art data sets.

Fig. 3
figure 3

Identification of different structural and individual distribution aspect

Fig. 4
figure 4

Identification of simple subset

4.1 Structural and individual distribution

The overall distribution has a huge impact in data quality. The SHLQI \(^2\) can be used to determine the overall structure due to homogeneity of clusters as mentioned above. As an example Fig. 3a visualizes the identification of homogeneous clusters. Herein, a classification data set was analyzed due to its data points being in a homogeneous cluster in which every data point has 60 direct nearest neighbors with the same class. The identification, therefore, needs to be a check if for a certain range of k the values of SHLQI \(^2\) for a point is not leaving the characteristic area of homogeneous clusters.

In contrast to analyzing a data set in terms of clustering it is possible to identify out of distributional data points individually. Out of distribution, therefore, needs to be defined as a data point being outside of its respective classes cluster but not interfering with another class. While interfering with another class, this data point would be a misclassificational outlier. The identification is quite similar to the one of homogeneous clusters. The only difference is, that due to average normalization of the input distances the value of QI \(^2\)R will be higher within a defined boundary of \(mlqi^2_{i,k} \epsilon [1|2]\). This is caused by considering significant higher normalized distances between the out of distribution data point and every other point inside the respective cluster in Eq. 5. Therefore, this identification is a check if for a certain local range of k the values of SHLQI \(^2\) for a point is within the range slightly above the homogeneous characteristic and below two as presented in Fig. 3b.

4.2 Locally simple input–output relationship

Data sets with multiple thousands of data points can be very complex when it comes to interpretation of the whole data set. Its input–output relationship behavior can have quite non-trivial changes or interruptions due to non-linearities in the real world. The data points can and will have heavily complex relationships especially with increasing input dimensions. Therefore, it is nice if a data set has a subset that can be specially treated due to a rather incomplex input–output relationship behavior. For those points, one can apply the divide-and-conquer principle. The simpler subset can be used to get further interpretability and stability into the system the data is applied on. For example, an ML system can generate high robustness against inputs near or inside this sub structure of the data. As an exemplary visualization of identification of simple subset, an artificially created data set representing a simple sine wave with heavy noise due to undefined reasons was created. The noisy parts follow no specific rules in terms of input–output correlation. In contrast to that, the sine is following a more strict relationship between input and output dimensions. The more linear this relationship is, the lower is the local complexity for the data points in those parts. A simple subset, therefore, will be visible due to persistent and falling low complexities in local areas. The longer the complexities are in a low area the more data points form a simple subset. Visually, this case is shown in Fig. 4.

It can be seen clearly that the marked lower complexities in the SHLQI \(^2\) directly refer to the sine based parts of the data set. Those data points are following defined simpler structural rules and can be treated like a simple subtask.

5 MNIST training data quality assurance

To explain the use of the new method, the popular MNIST handwritten digits training data set has been chosen. The data set consists of 60.000 different 28x28 gray-scale images. We applied the computation of SHLQI \(^2\) onto the whole data set with a maximum neighborhood of 3000 neighbors. At this point, the expectation was that nearly the whole histogram will be covered by the previous mentioned steep rises with extremely high value due to an average amount of 6000 data points per class. According to this amount of points per class, we expected the set to have a pretty decent clustering and, therefore, homogeneous groups with couple hundreds or thousands of neighbors from the same class. The next data point with another class would then cause an explosion of the value of \(mlqi^2_{i,k}\). As shown in Fig. 5, this assumption is totally met. The complexities of MNIST reached pretty high values with a peak at \(mlqi^2_{i,k}\approx 1100\) as well as high coverage of the histogram with classification characteristic steep rises in complexities.

Fig. 5
figure 5

SHLQI \(^2\) for MNIST training data with euclidean distance

5.1 Structural distribution

After first overall quality assurance, a deeper analysis to determine the structural distribution can be done. To get a good knowledge of how well the data are homogeneously clustered the proposed identification of structural distribution was applied. Figure 6 shows the correlation between the proposed identification and its respective black marked data points.

Table 1 Data points in homogeneous clusters due to neighborhood size with identical class

Interesting to see is that for the classes "one" (bottom left cluster), "zero" (rightmost cluster), and "two" (left cluster below "zero"), almost every data point is part of a homogeneous cluster with respect to 300 nearest neighbors. In contrast to that, the cluster "nine" has only two representative data points. This already shows the difference in quality assurance between SHLQI \(^2\) and dimension reduction algorithms. While UMAP represents a qualitative homogeneous cluster for class "nine", the SHLQI \(^2\) shows that it is build up by quantitative rather small homogeneous clusters within local areas. The UMAP visualization is misleading in terms of overall structural distribution. Furthermore, Table 1 shows the amount of data points building homogeneous cluster \(p_{cluster}\) with at least k nearest neighbors having an identical class.

Fig. 6
figure 6

Homogeneous clustering in MNIST training data with a neighborhood size of 300 nearest neighbors with identical class

5.2 Outlier detection

An outlier in classification tasks can be detected due to low MLQI \(^2\) at \(k=1\) and a steep rise in complexity for local areas \(k\epsilon [5|25]\). A low complexity at \(k=1\) represents data points having one direct neighbor with another class. This might as well be the case in terms of classification boundaries but in combination with steep complexity rises in local areas these classification boundaries will be filtered out. The filtering happens because of an evenly distributed amount of data points with another and the same class in local areas for boundaries. This would result in lower complexity values. As stated in above classification tasks have steep rises in complexity due to one or a few examples being in an homogeneous cluster which is the case for outliers. Examples for outliers, their class according to the data and their detection in MNIST training data regarding euclidean distance are shown in Fig. 7.

Fig. 7
figure 7

Outlier detection in MNIST data set

6 Conclusion

In this paper, we presented a novel quality assurance method based on visually representing the neighborhood relationship. We gave concrete examples for assurance of defined quality aspects and tested them exemplary with the MNIST training data. The QI \(^2\) can be used to get deeper structural and individual insights regarding qualitative and quantitative aspects of the data. The new tool-set based on QI \(^2\) enables quantitative requirements for the quality assurance process. As an example, a requirement for the quality assurance of the data set may be formulated, that for all potential outliers identified with SHLQI \(^2\) in neighborhoods of 10–50 examples, the output of every example has to be validated. With this measure, the probability of wrong class labels will be reduced significantly with relative low costs for quality assurance activities. The range for which neighborhoods the validation is required may be, e.g., defined according to the required safety integrity level of the application. Especially in combination with a representative visualization of the data, this tool has many handy features for an overall efficient interactive quality assurance process.