Skip to main content

Compare-xAI: Toward Unifying Functional Testing Methods for Post-hoc XAI Algorithms into a Multi-dimensional Benchmark

  • Conference paper
  • First Online:
Explainable Artificial Intelligence (xAI 2023)

Abstract

In recent years, Explainable AI (xAI) attracted a lot of attention as various countries turned explanations into a legal right. xAI algorithms enable humans to understand the underlying models and explain their behavior, leading to insights through which the models can be analyzed and improved beyond the accuracy metric by, e.g., debugging the learned pattern and reducing unwanted biases. However, the widespread use of xAI and the rapidly growing body of published research in xAI have brought new challenges. A large number of xAI algorithms can be overwhelming and make it difficult for practitioners to choose the correct xAI algorithm for their specific use case. This problem is further exacerbated by the different approaches used to assess novel xAI algorithms, making it difficult to compare them to existing methods. To address this problem, we introduce Compare-xAI, a benchmark that allows for a direct comparison of popular xAI algorithms with a variety of different use cases. We propose a scoring protocol employing a range of functional tests from the literature, each targeting a specific end-user requirement in explaining a model. To make the benchmark results easily accessible, we group the tests into four categories (fidelity, fragility, stability, and stress tests). We present results for 13 xAI algorithms based on 11 functional tests. After analyzing the findings, we derive potential solutions for data science practitioners as workarounds to the found practical limitations. Finally, Compare-xAI is a tentative to unify systematic evaluation and comparison methods for xAI algorithms with a focus on the end-user’s requirements. The code is made available at:

https://karim-53.github.io/cxai/.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 109.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 139.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Ribeiro, M.T., Singh, S., Guestrin, C.: Why should I trust you? Explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135–1144 (2016)

    Google Scholar 

  2. Dressel, J., Farid, H.: The accuracy, fairness, and limits of predicting recidivism. Sci. Advances 4(1), eaao5580 (2018)

    Google Scholar 

  3. Goodman, B., Flaxman, S.: European union regulations on algorithmic decision-making and a “right to explanation’’. AI Mag. 38(3), 50–57 (2017)

    Google Scholar 

  4. Sundararajan, M., Najmi, A.: The many Shapley values for model explanation. In: International Conference on Machine Learning, pp. 9269–9278. PMLR (2020)

    Google Scholar 

  5. Shapley, L.S.: Quota Solutions op N-person games1. Edited by Emil Artin and Marston Morse, p. 343 (1953)

    Google Scholar 

  6. Štrumbelj, E., Kononenko, I.: Explaining prediction models and individual predictions with feature contributions. Knowl. Inf. Syst. 41(3), 647–665 (2014)

    Article  Google Scholar 

  7. Lundberg, S.M., Lee, S.-I.: A unified approach to interpreting model predictions. In: Advances in Neural Information Processing Systems, vol. 30 (2017)

    Google Scholar 

  8. Sundararajan, M., Taly, A., Yan, Q.: Axiomatic attribution for deep networks. In: International Conference on Machine Learning, pp. 3319–3328. PMLR (2017)

    Google Scholar 

  9. Lundberg, S.M., Erion, G.G., Lee, S.-I.: Consistent individualized feature attribution for tree ensembles. arXiv preprint arXiv:1802.03888 (2018)

  10. Staniak, M., Biecek, P.: Explanations of model predictions with live and breakdown packages. arXiv preprint arXiv:1804.01955 (2018)

  11. Rozenblit, L., Keil, F.: The misunderstood limits of folk science: an illusion of explanatory depth. Cogn. Sci. 26(5), 521–562 (2002)

    Article  Google Scholar 

  12. Chromik, M., Eiband, M, Buchner, F., Krüger, A., Butz, A.: I think I get your point, AI! the illusion of explanatory depth in explainable AI. In: 26th International Conference on Intelligent User Interfaces, pp. 307–317 (2021)

    Google Scholar 

  13. Kaur, H., Nori, H., Jenkins, S., Caruana, R., Wallach, H., Vaughan, J.W.: Interpreting interpretability: understanding data scientists’ use of interpretability tools for machine learning. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–14 (2020)

    Google Scholar 

  14. Leavitt, M.L., Morcos, A.: Towards falsifiable interpretability research. arXiv preprint arXiv:2010.12016 (2020)

  15. Kearns, M., Roth, A.: The Ethical Algorithm: The Science of Socially Aware Algorithm Design. Oxford University Press, Oxford (2019)

    Google Scholar 

  16. Miller, T.: Explanation in artificial intelligence: insights from the social sciences. Artif. Intell. 267, 1–38 (2019)

    Article  MathSciNet  Google Scholar 

  17. Beizer, B.: Black-Box Testing: Techniques for Functional Testing of Software and Systems. Wiley, Hoboken (1995)

    Google Scholar 

  18. LeCun, Y., Cortes, C., Burges, C.J.: MNIST handwritten digit database. ATT Labs (2010). https://yann.lecun.com/exdb/mnist

  19. Covert, I., Lundberg, S.M., Lee, S.-I.: Understanding global feature contributions with additive importance measures. In: Advances in Neural Information Processing Systems, vol. 33, pp. 17212–17223 (2020)

    Google Scholar 

  20. Tsang, M., Rambhatla, S., Liu, Y.: How does this interaction affect me? Interpretable attribution for feature interactions. In: Advances in Neural Information Processing Systems, vol. 33, pp. 6147–6159 (2020)

    Google Scholar 

  21. Elizabeth Kumar, I., Venkatasubramanian, S., Scheidegger, C., Friedler, S.: Problems with Shapley-value-based explanations as feature importance measures. In: International Conference on Machine Learning, pp. 5491–5500. PMLR (2020)

    Google Scholar 

  22. Mohseni, S., Zarei, N., Ragan, E.D.: A multidisciplinary survey and framework for design and evaluation of explainable AI systems. arXiv preprint arXiv:1811.11839 (2018)

  23. Tsang, M., Enouen, J., Liu, Y.: Interpretable artificial intelligence through the lens of feature interaction. arXiv preprint arXiv:2103.03103 (2021)

  24. Angelov, P.P., Soares, E.A., Jiang, R., Arnold, N.I., Atkinson, P.M.: Explainable artificial intelligence: an analytical review. Wiley Interdisc. Rev. Data Min. Knowl. Discov. 11(5), e1424 (2021)

    Article  Google Scholar 

  25. Zhou, J., Gandomi, A.H., Chen, F., Holzinger, A.: Evaluating the quality of machine learning explanations: a survey on methods and metrics. Electronics 10(5), 593 (2021)

    Article  Google Scholar 

  26. Rozemberczki, B., et al.: The Shapley value in machine learning. arXiv preprint arXiv:2202.05594 (2022)

  27. Nauta, M., et al.: From anecdotal evidence to quantitative evaluation methods: a systematic review on evaluating explainable AI. arXiv preprint arXiv:2201.08164 (2022)

  28. Molnar, C.: Interpretable machine learning. Lulu.com (2020)

    Google Scholar 

  29. Sattarzadeh, S., Sudhakar, M., Plataniotis, K.N.: SVEA: a small-scale benchmark for validating the usability of post-hoc explainable AI solutions in image and signal recognition. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 4158–4167 (2021)

    Google Scholar 

  30. Greydanus, S.: Scaling down deep learning. arXiv preprint arXiv:2011.14439 (2020)

  31. DeYoung, J., et al.: Eraser: a benchmark to evaluate rationalized NLP models. arXiv preprint arXiv:1911.03429 (2019)

  32. Mohseni, S., Block, J.E., Ragan, E.D.: Quantitative evaluation of machine learning explanations: a human-grounded benchmark. arXiv preprint arXiv:1801.05075 (2020)

  33. Liu, Y., Khandagale, S., White, C., Neiswanger, W.: Synthetic benchmarks for scientific research in explainable machine learning. arXiv preprint arXiv:2106.12543 (2021)

  34. Sharma, A., Melnikov, V., Hüllermeier, E., Wehrheim, H.: Property-driven black-box testing of numeric functions. Softw. Eng. 2023 (2023)

    Google Scholar 

  35. Chrysostomou, G., Aletras, N.: Improving the faithfulness of attention-based explanations with task-specific information for text classification. arXiv preprint arXiv:2105.02657 (2021)

  36. Hamamoto, M., Egi, M.: Model-agnostic ensemble-based explanation correction leveraging Rashomon effect. In: 2021 IEEE Symposium Series on Computational Intelligence (SSCI), pp. 01–08. IEEE (2021)

    Google Scholar 

  37. Hameed, I., et al.: Based-XAI: breaking ablation studies down for explainable artificial intelligence. arXiv preprint arXiv:2207.05566 (2022)

  38. Guidotti, R.: Evaluating local explanation methods on ground truth. Artif. Intell. 291, 103428 (2021)

    Article  MathSciNet  Google Scholar 

  39. Lin, Y.-S., Lee, W.-C., Berkay Celik, Z.: What do you see? Evaluation of explainable artificial intelligence (XAI) interpretability through neural backdoors. In: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 1027–1035 (2021)

    Google Scholar 

  40. Hedström, A., et al.: An explainable AI toolkit for responsible evaluation of neural network explanations and beyond. J. Mach. Learn. Res. 24(34), 1–11 (2023)

    Google Scholar 

  41. Arras, L., Osman, A., Samek, W.: CLEVR-XAI: a benchmark dataset for the ground truth evaluation of neural network explanations. Inf. Fusion 81, 14–40 (2022)

    Article  Google Scholar 

  42. Du, M., Liu, N., Yang, F., Ji, S., Hu, X.: On attribution of recurrent neural network predictions via additive decomposition. In: The World Wide Web Conference, pp. 383–393 (2019)

    Google Scholar 

  43. Agarwal, C., et al.: OpenXAI: towards a transparent evaluation of model explanations. In: Advances in Neural Information Processing Systems, vol. 35, pp. 15784–15799 (2022)

    Google Scholar 

  44. Plumb, G., Molitor, D., Talwalkar, A.S.: Model agnostic supervised local explanations. In: Advances in Neural Information Processing Systems, vol. 31 (2018)

    Google Scholar 

  45. Owen, G.: Multilinear extensions of games. Manag. Sci. 18(5-part-2), 64–79 (1972)

    Google Scholar 

  46. Sundararajan, M., Dhamdhere, K., Agarwal, A.: The Shapley Taylor interaction index. In: International Conference on Machine Learning, pp. 9259–9268. PMLR (2020)

    Google Scholar 

  47. Ghorbani, A., Abid, A., Zou, J.: Interpretation of neural networks is fragile. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 3681–3688 (2019)

    Google Scholar 

  48. Janzing, D., Minorics, L., Blöbaum, P.: Feature relevance quantification in explainable AI: a causal problem. In: International Conference on Artificial Intelligence and Statistics, pp. 2907–2916. PMLR (2020)

    Google Scholar 

  49. Hooker, G., Mentch, L., Zhou, S.: Unrestricted permutation forces extrapolation: variable importance requires at least one more model, or there is no free variable importance. Stat. Comput. 31(6), 1–16 (2021)

    Article  MathSciNet  Google Scholar 

  50. Lakkaraju, H., Kamar, E, Caruana, R., Leskovec, J.: Faithful and customizable explanations of black box models. In: Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, pp. 131–138 (2019)

    Google Scholar 

  51. Melis, D.A., Jaakkola, T.: Towards robust interpretability with self-explaining neural networks. In: Advances in Neural Information Processing Systems, vol. 31 (2018)

    Google Scholar 

  52. Lakkaraju, H., Bastani, O.: “How do I fool you?” manipulating user trust via misleading black box explanations. In: Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, pp. 79–85 (2020)

    Google Scholar 

  53. Mollas, I., Bassiliades, N., Tsoumakas, G.: Altruist: argumentative explanations through local interpretations of predictive models. In: Proceedings of the 12th Hellenic Conference on Artificial Intelligence, pp. 1–10 (2022)

    Google Scholar 

  54. Larson, J., Mattu, S., Kirchner, L., Angwin, J.: How we analyzed the COMPAS recidivism algorithm. ProPublica 9(1), 3 (2016)

    Google Scholar 

  55. Jeyakumar, J.V., Noor, J., Cheng, Y.-H., Garcia, L., Srivastava, M.: How can I explain this to you? An empirical study of deep neural network explanation methods. In: Advances in Neural Information Processing Systems, vol. 33, pp. 4211–4222 (2020)

    Google Scholar 

  56. Harris, C., Pymar, R., Rowat, C.: Joint Shapley values: a measure of joint feature importance. arXiv preprint arXiv:2107.11357 (2021)

Download references

Acknowledgments and Disclosure of Funding

The authors gratefully acknowledge the support and funding provided by IDIADA Fahrzeugtechnik GmbH, Munich, Germany and Dr. Ing. h.c. F. Porsche AG, Stuttgart, Germany. Their generous contribution enabled the successful completion of this research project.

We thank Dorra El Mekki, Maximilian Muschalik, Patrick Kolpaczki and Michael Rapp for their thoughtful feedback on earlier iterations of this work.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mohamed Karim Belaid .

Editor information

Editors and Affiliations

1 Electronic supplementary material

Below is the link to the electronic supplementary material.

Supplementary material 1 (pdf 132 KB)

Appendices

A Tests

For the proof-of-concept, the following list of tests is considered. Note that some tests count twice as they test both feature importance and feature attribution.

cough_and_fever:

answers the following question: Can the xAI algorithm detect symmetric binary input features?. The trained model’s equation is [Cough AND Fever]*80. The test utilize XGBRegressor model trained on a synthetic uniform distribution dataset (total size: 20000). The test procedure is as follows: train a model such that its response to the two features is exactly the same. The xAI algorithm should detect symmetric features (equal values) and allocate them equal importance. The score is calculated as follows: 1 if the xAI detect the two features are symmetric. 0 if the difference in importance is above one unit. The test is classified in the fidelity category because it is a simple tree model that demonstrate inconsistencies in explanation [9].

cough_and_fever_10_90:

answers the following question: Can the xAI algorithm detect that ’Cough’ feature is more important than ’Fever’?. The trained model’s equation is [Cough AND Fever]*80 + [Cough]*10. Cough should be more important than Fever globally. Locally for the case (Fever = yes, Cough = yes) the feature attribution of Cough should be more important. The test utilize XGBRegressor model trained on a synthetic uniform distribution dataset (total size: 20000). The test procedure is as follows: train a model with two features with unequal impact on the model. The feature with a higher influence on the output should be detected more important. The score is calculated as follows: Return 1 if Cough is more important otherwise 0. The test is classified in the fidelity category because it is a simple tree model that demonstrate inconsistencies in explanation due to the tree structure [9].

x0_plus_x1_distrib_non_uniform_stat_indep:

answers the following question: Is the xAI able to explain the model correctly despite a non-uniform distribution of the data?. The test demonstrate the effect of data distribution/causal inference. The test utilize XGBRegressor model trained on a non-uniform and statistically independent dataset (total size: 10000). The test procedure is as follows: Check if the explanation change when the distribution change. Check if non-uniform distributions affect the explanation. The score is calculated as follows: returns 1 if the two binary features obtain the same importance. The test is classified in the stability category because it assesses the impact of slightly changing the inputs [48].

x0_plus_x1_distrib_uniform_stat_dep:

answers the following question: Is the xAI able to explain the model correctly despite a statistically-dependent distribution of the data?. The test demonstrate the effect of data distribution/causal inference. The example was given in both [49] and [48]. The test utilize XGBRegressor model trained on a uniform and statistically dependent dataset (total size: 10000). The test procedure is as follows: Check if the explanation change when the distribution change. Check if statistically dependent distributions affect the explanation. The score is calculated as follows: returns 1 if the two binary features obtain the same importance. The test is classified in the stability category because To assess the impact of changing the inputs of f... This way, we are able to talk about a hypothetical scenario where the inputs are changed compared to the true features [48].

mnist:

answers the following question: Is the xAI able to detect all dummy (constant and useless) pixels?. The xAI algorithm should detect that important pixels are only in the center of the image. The test utilize an MLP model trained on the MNIST dataset (total size: 70000). The test procedure is as follows: simply train and explain the MLP model globally for every pixel. The score is calculated as follows: Return the ratio of constant pixels detected as dummy divided by the true number of constant pixels. The test is classified in the stress category because of the high number of input features. The test is adapted from [19].

fooling_perturbation_alg:

answers the following question: Is the xAI affected by an adversarial attack against perturbation-based algorithms?. Model-agnostic xAI algorithms that use feature perturbation methods might be vulnerable to this attack. The adversarial attack exploits a vulnerability to lower the feature importance of a specific feature. Setup: Let’s begin by examining the COMPAS data set. This data set consists of defendant information from Broward County, Florida. Let’s suppose that some adversary wants to mask biased or racist behavior on this data set. The test utilizes a custom function model trained on the COMPAS dataset (total size: 4629). The test procedure is as follows: The xAI algorithms need to explain the following corrupted model (custom function): if the input is from the dataset then the output is from a biased model. if not then the output is from a fair model. The score is calculated as follows: Return 1 if Race is the most important feature despite the adversarial attack. The score decreases while its rank decrease. The test is classified in the fragility category because fragility includes all adversarial attacks [47].

counterexample_dummy_axiom:

answers the following question: Is the xAI able to detect unused input features?. This is a counter-example used in literature to verify that SHAP CES do not satisfy the dummy axiom while BSHAP succeeds in this test. The test utilizes a custom function model trained on a synthetic dataset (total size: 20000). The test procedure is as follows: Train a model with one extra feature B that is dummy. The score is calculated as follows: returns 1 if the dummy feature B obtains a null importance. The test is classified in the fidelity category because assigning an importance of zero to a dummy feature reflects the model behavior (Fidelity) but also helps the data scientist to quickly understand the model.

a_and_b_or_c:

answers the following question: Can the xAI algorithm detect that input feature ’A’ is more important than ’B’ or ’C’?. This is a baseline test that the xAI should succeed in all cases. Model: A and (B or C). Goal: make sure that A is more important than B, C. Noise effect: even if the model output is not exactly equal to 1 still we expect the xai to give a correct answer. The test utilize XGBRegressor model trained on a synthetic dataset (total size: 20000). The test procedure is as follows: The model learns the following equation: A and (B or C). The explanation should prove that A is more important. The score is calculated as follows: If A is the most important feature then return 1. If A is the 2nd most important feature then return 0.5 i.e. 1- (1/nb of feature more important than A). If A is the last one: return 0 (completely wrong). The test is classified in the fidelity category because of the same reason as cough and fever 10–90: A’s effect on the output is higher than B or C.

correlated_features:

answers the following question: Can the xAI algorithm detect, that two of three input features are perfectly correlated?. This is a fragility test, which attacks the xAI algorithm by introducing a third feature ’C’ which is perfectly correlated with feature ’B’. The test utilizes a XGBRegressor model trained on a synthetic dataset with three input features (total size: 20000). The test procedure is as follows: The model is trained on a synthetic dataset with the three input features ’A’, ’B’ and ’C’, of which ’B’ and ’C’ are perfectly correlated. The xAI algorithm should detect this correlation and only assign non-zero importance to one of the two correlated features. The score is calculated as follows: Since we do not care about which correlated feature is assigned the non zero importance value, we emply following scoring metric: 1-(min(B, C)/max(B, C)). This function assigns a value of zero if the assigned importance of the correlated features is equal, and 1 if one of the features is assigned an importance of zero. The test is classified in the fragility category because it tests the xAI algorithms ability to adjust to perfect correlation in the input features.

B xAI Algorithms

archipelago:

 [20] separate the input features into sets. all features inside a set interact and there is no interaction outside a set. ArchAttribute is an interaction attribution method. ArchDetect is the corresponding interaction detector. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature interaction (local explanation).

baseline_random:

 [33] Output a random explanation. It is not a real explainer. It helps measure the baseline score and processing time. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature attribution (local explanation), Feature importance (global explanation), Feature interaction (local explanation).

exact_shapley_values:

 [5] is a permutation-based xAI algorithm following a game theory approach: Iteratively Order the features randomly, then add them to the input one at a time following this order, and calculate their expected marginal contribution [4]. The output is unique given a set of constrains defined in the original paper. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature importance (global explanation). The following information are required by the xAI algorithm:, A reference dataset (input only), The model’s predict function

kernel_shap:

 [7] it approximates the Shapley values with a constant noise [48]. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature attribution (local explanation), Feature importance (global explanation). The following information are required by the xAI algorithm:, A reference dataset (input only), The model’s predict function

lime:

 [1] it explains the model locally by generating an interpretable model approximating the original one. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature attribution (local explanation), Feature importance (global explanation). The following information are required by the xAI algorithm:, A reference dataset (input only), The model’s predict probability function, Nature of the ML task (regression/classification), The model’s predict function

maple:

 [44] is a supervised neighborhood approach that combines ideas from local linear models and ensembles of decision trees [44]. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature attribution (local explanation), Feature importance (global explanation). The following information are required by the xAI algorithm:, AI model’s structure, A reference dataset (input only), The train set, The model’s predict function

partition:

 [7] Partition SHAP approximates the Shapley values using a hierarchy of feature coalitions. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature attribution (local explanation), Feature importance (global explanation). The following information are required by the xAI algorithm:, A reference dataset (input only), The model’s predict function

permutation:

is a shuffle-based feature importance. It permutes the input data and compares it to the normal prediction The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature attribution (local explanation), Feature importance (global explanation). The following information are required by the xAI algorithm:, input features, A reference dataset (input only), The model’s predict function

permutation_partition:

is a combination of permutation and partition algorithm from shap. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature attribution (local explanation), Feature importance (global explanation). The following information are required by the xAI algorithm:, input features, A reference dataset (input only), The model’s predict function

saabas:

explain tree based models by decomposing each prediction into bias and feature contribution components The xAI algorithm can explain tree-based models. The xAI algorithm can output the following explanations: Feature attribution (local explanation), Feature importance (global explanation). The following information are required by the xAI algorithm:, AI model’s structure

sage:

 [19] Compute feature importance based on Shapley value but faster. The features that are most critical for the model to make good predictions will have large importance and only features that make the model’s performance worse will have negative values.

Disadvantage: The convergence of the algorithm depends on 2 parameters: ‘thres‘ and ‘gap‘. The algorithm can be trapped in a potential infinite loop if we do not fine tune them. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature importance (global explanation). The following information are required by the xAI algorithm:, True output of the data points to explain, A reference dataset (input only), The model’s predict function

shap_interaction:

 [45] SI: Shapley Interaction Index. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature interaction (local explanation).

shapley_taylor_interaction:

 [46] STI: Shapley Taylor Interaction Index. The xAI algorithm is model agnostic i.e. it can explain any AI model. The xAI algorithm can output the following explanations: Feature interaction (local explanation).

tree_shap:

 [9] accurately compute the shap values using the structure of the tree model. The xAI algorithm can explain tree-based models. The xAI algorithm can output the following explanations: Feature attribution (local explanation), Feature importance (global explanation). The following information are required by the xAI algorithm:, AI model’s structure, A reference dataset (input only)

tree_shap_approximation:

is a faster implementation of shap reserved for tree based models. The xAI algorithm can explain tree-based models. The xAI algorithm can output the following explanations: Feature attribution (local explanation), Feature importance (global explanation). The following information are required by the xAI algorithm:, AI model’s structure, A reference dataset (input only)

joint_shapley:

is an extension of the axioms and intuitions of Shapley values proposed by [56]. This xAI algorithm creates a powerset up to a user specified power k of all the input features and computes the average Shapley values for all different subsets. This leads to a more accurate attribution of importance for each input feature, but significantly increases the run-time.

C Test Results

Table 2 contains test results without using any filter. Tests is the number of completed tests. Time is the average execution time per test. It informs the user about the relative difference in execution time between algorithms.

Table 2. Results for all Tests

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Belaid, M.K., Bornemann, R., Rabus, M., Krestel, R., Hüllermeier, E. (2023). Compare-xAI: Toward Unifying Functional Testing Methods for Post-hoc XAI Algorithms into a Multi-dimensional Benchmark. In: Longo, L. (eds) Explainable Artificial Intelligence. xAI 2023. Communications in Computer and Information Science, vol 1902. Springer, Cham. https://doi.org/10.1007/978-3-031-44067-0_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-44067-0_5

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-44066-3

  • Online ISBN: 978-3-031-44067-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics