Abstract
Artificial Intelligence (AI) systems have grown commonplace in modern life, with various applications from customized suggestions to self-driving vehicles. As these systems get more complicated, the necessity for transparency in their decision-making processes becomes more critical. Explainability refers to an AI system’s ability to explain how and why it made a certain judgement or prediction. Recently, there has been a surge of interest in constructing explainable AI (XAI) systems that give insights into the decision-making processes of machine learning models. This paper discloses and elaborates upon a selection of XAI techniques, identifies current challenges and possible future directions in XAI research.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Ambhaikar, A.: A survey on health care and expert system. Math. Statist. Eng. Appl. 72(1), 451–461 (2023)
Bahani, K., Moujabbir, M., Ramdani, M.: An accurate fuzzy rule-based classification systems for heart disease diagnosis. Sci. African 14, e01019 (2021)
Baydin, A.G., Pearlmutter, B.A., Radul, A.A., Siskind, J.M.: Automatic differentiation in machine learning: a survey. J. Mach. Learn. Res. 18, 1–43 (2018)
Burkhardt, S., Brugger, J., Wagner, N., Ahmadi, Z., Kersting, K., Kramer, S.: Rule extraction from binary neural networks with convolutional rules for model validation. Front. Artif. Intell. 4, 642263 (2021)
Cambra Baseca, C., Sendra, S., Lloret, J., Tomas, J.: A smart decision system for digital farming. Agronomy 9(5), 216 (2019)
Choraś, M., Pawlicki, M., Puchalski, D., Kozik, R.: Machine learning–the results are not the only thing that matters! what about security, explainability and fair- ness? In: Krzhizhanovskaya, V.V. et al (eds.). Computational Science – ICCS 2020. ICCS 2020. Lecture Notes in Computer Science, vol. 12140, pp. 615–628. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-50423-6_46
Domingos, P.: A few useful things to know about machine learning. Commun. ACM 55(10), 78–87 (2012)
Doshi-Velez, F., Kim, B.: Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608 (2017)
Dwivedi, R., et al.: Explainable ai (xai): Core ideas, techniques, and solutions. ACM Comput. Surv. 55(9), 1–33 (2023)
Friedman, J.H., Popescu, B.E.: Predictive learning via rule ensembles. Annal. Appl. Statist. 2, 916–954 (2008)
Goodfellow, I., Bengio, Y., Courville, A.: Deep learning. MIT press (2016)
Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., Pedreschi, D.: A survey of methods for explaining black box models. ACM computing surveys (CSUR) 51(5), 1–42 (2018)
Han, J., Kamber, M., Pei, J.: Data mining concepts and techniques third edition. University of Illinois at Urbana-Champaign Micheline Kamber Jian Pei Simon Fraser University (2012)
Henderson, J., et al.: Certifai: a toolkit for building trust in AI systems. In: Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, pp. 5249–5251 (2021)
Liao, Q.V., Gruen, D., Miller, S.: Questioning the AI: informing design practices for explainable AI user experiences. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–15 (2020)
Liao, Q.V., Varshney, K.R.: Human-centered explainable ai (xai): From algorithms to user experiences. arXiv preprint arXiv:2110.10790 (2021)
Lipton, Z.C.: The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery. Queue 16(3), 31–57 (2018)
Luo, C., et al.: Rulefit-based nomogram using inflammatory indicators for predicting survival in nasopharyngeal carcinoma, a bi-center study. J. Inflamm. Res. 15, 4803–4815 (2022)
Mahbooba, B., Timilsina, M., Sahal, R., Serrano, M.: Explainable artificial intelligence (xai) to enhance trust management in intrusion detection systems using decision tree model. Complexity 2021, 1–11 (2021)
Mitrović, S., Andreoletti, D., Ayoub, O.: Chatgpt or human? detect and explain. Explaining Decisions of Machine Learning Model for Detecting Short Chatgpt- Generated Text. arXiv preprint arXiv:2301.13852 (2023)
Molnar, C.: Interpretable machine learning. Lulu.com (2020)
Nalepa, G., Araszkiewicz, M., Nowaczyk, S., Bobek, S.: Building trust to AI systems through explainability: technical and legal perspectives (2019)
Nwakanma, C.I., et al.: Explainable artificial intelligence (xai) for intrusion detection and mitigation in intelligent connected vehicles: a review. Appl. Sci. 13(3), 1252 (2023)
Panesar, A.: Machine learning and AI for healthcare. Springer (2019). https://doi.org/10.1007/978-1-4842-3799-1
Quinlan, J.R.: Induction of decision trees. Machine learning 1, 81–106 (1986)
Reddy, B., Fields, R.: From past to present: a comprehensive technical review of rule-based expert systems from 1980–2021. In: Proceedings of the 2022 ACMSoutheast Conference, pp. 167–172 (2022)
Ribeiro, M.T., Singh, S., Guestrin, C.: “Why should I trust you?” explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135–1144 (2016)
Roth, A.M., Liang, J., Manocha, D.: Xai-n: Sensor-based robot navigation using expert policies and decision trees. In: 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2053–2060. IEEE (2021)
Samek, W., Wiegand, T., Müller, K.R.: Explainable artificial intelligence: understanding, visualizing and interpreting deep learning models. arXiv preprint arXiv:1708.08296 (2017)
Schaaf, N., Huber, M., Maucher, J.: Enhancing decision tree based interpretation of deep neural networks through l1-orthogonal regularization. In: 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA), pp. 42–49. IEEE (2019)
Sharma, S., Henderson, J., Ghosh, J.: Certifai: counterfactual explanations for robustness, transparency, interpretability, and fairness of artificial intelligence models. arXiv preprint arXiv:1905.07857 (2019)
Szczepański, M., Choraś, M., Pawlicki, M., Pawlicka, A.: The methods and approaches of explainable artificial intelligence. In: Paszynski, M., Kranzlmüller, D., Krzhizhanovskaya, V.V., Dongarra, J.J., Sloot, P.M. (eds.) Computational Science – ICCS 2021. ICCS 2021. Lecture Notes in Computer Science, vol. 12745, pp. 3–17. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-77970-2_1
Szczepański, M., Pawlicki, M., Kozik, R., Choraś, M.: New explainability method for bert-based model in fake news detection. Sci. Rep. 11(1), 23705 (2021)
Van der Velden, B.H., Kuijf, H.J., Gilhuijs, K.G., Viergever, M.A.: Explainable artificial intelligence (xai) in deep learning-based medical image analysis. Med. Image Anal. 79, 102470 (2022)
Vouros, G.A.: Explainable deep reinforcement learning: state of the art and challenges. ACM Comput. Surv. 55(5), 1–39 (2022)
Zhang, Z., Hamadi, H.A., Damiani, E., Yeun, C.Y., Taher, F.: Explainable artificial intelligence applications in cyber security: State-of-the-art in research. arXiv preprint arXiv:2208.14937 (2022)
Acknowledgements
This work is funded under the AI4Cyber project, which has received funding from the European Union’s Horizon Europe research and innovation programme under grant agreement No 101070450.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Kurek, W., Pawlicki, M., Pawlicka, A., Kozik, R., Choraś, M. (2023). Explainable Artificial Intelligence 101: Techniques, Applications and Challenges. In: Huang, DS., Premaratne, P., Jin, B., Qu, B., Jo, KH., Hussain, A. (eds) Advanced Intelligent Computing Technology and Applications. ICIC 2023. Lecture Notes in Computer Science(), vol 14089. Springer, Singapore. https://doi.org/10.1007/978-981-99-4752-2_26
Download citation
DOI: https://doi.org/10.1007/978-981-99-4752-2_26
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-4751-5
Online ISBN: 978-981-99-4752-2
eBook Packages: Computer ScienceComputer Science (R0)