Skip to main content

Explain to Not Forget: Defending Against Catastrophic Forgetting with XAI

  • Conference paper
  • First Online:
Machine Learning and Knowledge Extraction (CD-MAKE 2022)

Abstract

The ability to continuously process and retain new information like we do naturally as humans is a feat that is highly sought after when training neural networks. Unfortunately, the traditional optimization algorithms often require large amounts of data available during training time and updates w.r.t. new data are difficult after the training process has been completed. In fact, when new data or tasks arise, previous progress may be lost as neural networks are prone to catastrophic forgetting. Catastrophic forgetting describes the phenomenon when a neural network completely forgets previous knowledge when given new information. We propose a novel training algorithm called Relevance-based Neural Freezing in which we leverage Layer-wise Relevance Propagation in order to retain the information a neural network has already learned in previous tasks when training on new data. The method is evaluated on a range of benchmark datasets as well as more complex data. Our method not only successfully retains the knowledge of old tasks within the neural networks but does so more resource-efficiently than other state-of-the-art solutions.

S. Ede and S. Baghdadlian—Contributed equally.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Alber, M., et al.: iNNvestigate neural networks! J. Mach. Learn. Res. 20(93), 1–8 (2019)

    MathSciNet  Google Scholar 

  2. Anders, C.J., Neumann, D., Samek, W., Müller, K.-R., Lapuschkin, S.: Software for dataset-wide XAI: from local explanations to global insights with Zennit, CoRelAy, and ViRelAy. arXiv preprint arXiv:2106.13200 (2021)

  3. Bach, S., Binder, A., Montavon, G., Klauschen, F., Müller, K.-R., Samek, W.: On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS ONE 10(7), e0130140 (2015)

    Article  Google Scholar 

  4. Baehrens, D., Schroeter, T., Harmeling, S., Kawanabe, M., Hansen, K., Müller, K.-R.: How to explain individual classification decisions (2010)

    Google Scholar 

  5. Becking, D., Dreyer, M., Samek, W., Müller, K., Lapuschkin, S.: ECQ\(^\text{ x }\): explainability-driven quantization for low-bit and sparse DNNs. In: Holzinger, A., Goebel, R., Fong, R., Moon, T., Müller, K.R., Samek, W. (eds.) xxAI 2020. LNCS, vol. 13200, pp. 271–296. Springer, Cham (2022). https://doi.org/10.1007/978-3-031-04083-2_14

    Chapter  Google Scholar 

  6. Chereda, H., et al.: Explaining decisions of graph convolutional neural networks: patient-specific molecular subnetworks responsible for metastasis prediction in breast cancer. Genome Med. 13(1), 1–16 (2021). https://doi.org/10.1186/s13073-021-00845-7

    Article  Google Scholar 

  7. Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255 (2009)

    Google Scholar 

  8. Deng, L.: The MNIST database of handwritten digit images for machine learning research. IEEE Sig. Process. Mag. 29(6), 141–142 (2012)

    Article  Google Scholar 

  9. Eidinger, E., Enbar, R., Hassner, T.: Age and gender estimation of unfiltered faces. IEEE Trans. Inf. Forensics Secur. 9(12), 2170–2179 (2014)

    Article  Google Scholar 

  10. Erhan, D., Bengio, Y., Courville, A., Vincent, P.: Visualizing higher-layer features of a deep network. Technical report, Univeristé de Montréal, January 2009

    Google Scholar 

  11. Evans, T., et al.: The explainability paradox: challenges for xAI in digital pathology. Future Gener. Comput. Syst. 133, 281–296 (2022)

    Article  Google Scholar 

  12. Farquhar, S., Gal, Y.: Towards robust evaluations of continual learning. arXiv preprint arXiv:1805.09733 (2018)

  13. Fong, R.C., Vedaldi, A.: Interpretable explanations of black boxes by meaningful perturbation. In: 2017 IEEE International Conference on Computer Vision (ICCV) (2017)

    Google Scholar 

  14. French, R.: Catastrophic forgetting in connectionist networks. Trends Cogn. Sci. 3, 128–135 (1999)

    Article  Google Scholar 

  15. Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003)

    MATH  Google Scholar 

  16. Guyon, I., Weston, J., Barnhill, S., Vapnik, V.: Gene selection for cancer classification using support vector machines. Mach. Learn. 46, 389–422 (2002). https://doi.org/10.1023/A:1012487302797

    Article  MATH  Google Scholar 

  17. Hohman, F., Park, H., Robinson, C., Chau, D.H.: Summit: scaling deep learning interpretability by visualizing activation and attribution summarizations. arXiv preprint arXiv:1904.02323 (2019)

  18. Hägele, M., et al.: Resolving challenges in deep learning-based analyses of histopathological images using explanation methods. Sci. Rep. 10, 6423 (2020)

    Article  Google Scholar 

  19. Kim, B., et al.: Interpretability beyond feature attribution: quantitative testing with concept activation vectors (TCAV). In: International Conference on Machine Learning, pp. 2668–2677. PMLR (2018)

    Google Scholar 

  20. Kirkpatrick, J., et al.: Overcoming catastrophic forgetting in neural networks (2017)

    Google Scholar 

  21. Kohlbrenner, M., Bauer, A., Nakajima, S., Binder, A., Samek, W., Lapuschkin, S.: Towards best practice in explaining neural network decisions with LRP. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–7 (2020)

    Google Scholar 

  22. Krizhevsky, A.: Learning multiple layers of features from tiny images. Master’s thesis, University of Toronto, Department of Computer Science (2009)

    Google Scholar 

  23. Lange, M.D., et al.: Continual learning: a comparative study on how to defy forgetting in classification tasks. arXiv preprint arXiv:1909.08383 (2019)

  24. Lee, J., Yoon, J., Yang, E., Hwang, S.J.: Lifelong learning with dynamically expandable networks. arXiv preprint arXiv:1708.01547 (2017)

  25. Montavon, G., Lapuschkin, S., Binder, A., Samek, W., Müller, K.-R.: Explaining nonlinear classification decisions with deep Taylor decomposition. Pattern Recogn. 65, 211–222 (2017)

    Article  Google Scholar 

  26. Nguyen, A.M., Dosovitskiy, A., Yosinski, J., Brox, T., Clune, J.: Synthesizing the preferred inputs for neurons in neural networks via deep generator networks. arXiv preprint arXiv:1605.09304 (2016)

  27. Olah, C., Mordvintsev, A., Schubert, L.: Feature visualization. Distill 2(11), e7 (2017)

    Article  Google Scholar 

  28. Oren, G., Wolf, L.: In defense of the learning without forgetting for task incremental learning. In: Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, pp. 2209–2218 (2021)

    Google Scholar 

  29. Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., et al.: PyTorch: an imperative style, high-performance deep learning library. In: Advances in Neural Information Processing Systems, pp. 8024–8035 (2019)

    Google Scholar 

  30. Radford, A., et al.: Learning transferable visual models from natural language supervision (2021)

    Google Scholar 

  31. Ribeiro, M.T., Singh, S., Guestrin, C.: “Why should I trust you?”: Explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135–1144 (2016)

    Google Scholar 

  32. Samek, W., Binder, A., Montavon, G., Bach, S., Müller, K.-R.: Evaluating the visualization of what a deep neural network has learned. IEEE Trans. Neural Netw. Learn. Syst. 28(11), 2660–2673 (2017)

    Article  MathSciNet  Google Scholar 

  33. Samek, W., Montavon, G., Lapuschkin, S., Anders, C.J., Müller, K.-R.: Explaining deep neural networks and beyond: a review of methods and applications. Proc. IEEE 109(3), 247–278 (2021)

    Article  Google Scholar 

  34. Samek, W., Wiegand, T., Müller, K.-R.: Explainable artificial intelligence: understanding, visualizing and interpreting deep learning models. ITU J. ICT Discov. 1(1), 39–48 (2018)

    Google Scholar 

  35. Schuhmann, C., et al.: LAION-400M: open dataset of clip-filtered 400 million image-text pairs. arXiv preprint arXiv:2111.02114 (2021)

  36. Serrà, J., Surís, D., Miron, M., Karatzoglou, A.: Overcoming catastrophic forgetting with hard attention to the task. arXiv preprint arXiv:1801.01423 (2018)

  37. Silver, D., Huang, A., Maddison, C., Guez, A., Sifre, L., Driessche, G., et al.: Mastering the game of go with deep neural networks and tree search. Nature 529, 484–489 (2016)

    Article  Google Scholar 

  38. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: 3rd International Conference on Learning Representations (ICLR) (2015)

    Google Scholar 

  39. Sundararajan, M., Taly, A., Yan, Q.: Axiomatic attribution for deep networks. In: International Conference on Machine Learning, pp. 3319–3328. PMLR (2017)

    Google Scholar 

  40. van de Ven, G.M., Tolias, A.S.: Three scenarios for continual learning. arXiv preprint arXiv:1904.07734 (2019)

  41. Wilm, F., Benz, M., Bruns, V., Baghdadlian, S., Dexl, J., Hartmann, D., et al.: Fast whole-slide cartography in colon cancer histology using superpixels and CNN classification. J. Med. Imaging 9(2), 027501 (2022)

    Article  Google Scholar 

  42. Wortsman, M., et al.: Supermasks in superposition. arXiv preprint arXiv:2006.14769 (2020)

  43. Wu, Y., et al.: Large scale incremental learning. arXiv preprint arXiv:1905.13260 (2019)

  44. Yeom, S.K., et al.: Pruning by explaining: a novel criterion for deep neural network pruning. Pattern Recogn. 115, 107899 (2021)

    Article  Google Scholar 

  45. Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. arXiv preprint arXiv:1311.2901 (2013)

  46. Zenke, F., Poole, B., Ganguli, S.: Improved multitask learning through synaptic intelligence. arXiv preprint arXiv:1703.04200 (2017)

  47. Zintgraf, L.M., Cohen, T.S., Adel, T., Welling, M.: Visualizing deep neural network decisions: prediction difference analysis. arXiv preprint arXiv:1702.04595 (2017)

Download references

Acknowledgment

This work was supported by the German Ministry for Education and Research as BIFOLD (ref. 01IS18025A and ref. 01IS18037A), the European Union’s Horizon 2020 programme (grant no. 965221 and 957059), and the Investitionsbank Berlin under contract No. 10174498 (Pro FIT programme).

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Wojciech Samek or Sebastian Lapuschkin .

Editor information

Editors and Affiliations

A Appendix

A Appendix

1.1 A.1 MNIST-Split

The model architecture was taken from [40], which compares multiple methods for mitigating catastrophic forgetting. It consists of two hidden layers with 400 neurons each and ReLU activations. As for most experiments except the real-world dataset, the pruning threshold is set to 2%, meaning that the accuracy can drop by up to 2% before the pruning procedure is halted. We use the Adam optimizer with a learning rate of 0.001, \(\beta _1 = 0.9\) and \(\beta _2=0.999\) with a batch size of 128.

1.2 A.2 MNIST-Permuted

For this experiment, the architecture from [34] was adapted by increasing the number of hidden layer units to 1000 per layer to match the increased complexity of the task. Additionally, the learning rate was decreased to 0.0001 and the model was trained for ten instead of four epochs per task.

1.3 A.3 CIFAR10 and CIFAR100

In this experiment, we adopted architecture and experimental setup from [46].

1.4 A.4 ImageNet Split

Here, we replicate the conditions from [43] but establish our baseline after ten instead of 70 epochs, which we also use when applying RNF.

1.5 A.5 Adience

As is state-of-the art for this dataset [9], we normalize to zero mean and unit standard deviation during training and apply data augmentation for the training data by randomly cropping to 224 \(\times \) 224 as well as horizontal flipping. For testing, each sample is cropped five times to 224 \(\times \) 224 (four corner crops and one center crop), where each crop is additionally mirrored horizontally. The ground truth is then compared to the mean of the Softmax activations of the ten samples. Only the center crops are used in the reference data. As the dataset is strongly imbalanced, we additionally employ a resampling strategy during training that undersamples classes with a high number of samples and oversamples classes with a low number of samples by computing the class probabilities and then sampling from a multinomial distribution.

In this experiment, we employ a VGG-16 network architecture [38] that has been pretrained on ImageNet (from the PyTorch [29] model zoo), as well as an Adam optimizer and L2 regularization.

  • Split: We used a learning rate of 0.0001, L2 regularization with \(\lambda = 0.01\) and a batch size of 32.

  • Entire Dataset: The model was trained with a learning rate of 0.00001 and \(\lambda = 0.001\).

Rights and permissions

Reprints and permissions

Copyright information

© 2022 IFIP International Federation for Information Processing

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ede, S. et al. (2022). Explain to Not Forget: Defending Against Catastrophic Forgetting with XAI. In: Holzinger, A., Kieseberg, P., Tjoa, A.M., Weippl, E. (eds) Machine Learning and Knowledge Extraction. CD-MAKE 2022. Lecture Notes in Computer Science, vol 13480. Springer, Cham. https://doi.org/10.1007/978-3-031-14463-9_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-14463-9_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-14462-2

  • Online ISBN: 978-3-031-14463-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics