Skip to main content
  • 2915 Accesses

Abstract

Post-hoc techniques represent a vast collection of methods created to specifically address the black-box problem, where we do not have access to the internal feature representations or model structure. There are considerable advantages to using post-hoc methods. They can work for a wide variety of model algorithms. They allow for different representations to be used for internal modeling and explanation. They can also provide different types of explanations for the same model. However, there is a trade-off between the fidelity and comprehensibility of explanations.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 119.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 159.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 159.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. A. Fisher, C. Rudin, F. Dominici, All models are wrong, but many are useful: learning a variable’s importance by studying an entire class of prediction models simultaneously (2019). arXiv:1801.01489 [stat.ME]

    Google Scholar 

  2. J.H. Friedman, B.E. Popescu, Predictive learning via rule ensembles. Ann. Appl. Stat. 2(3) (2008). ISSN: 1932-6157. http://dx.doi.org/10.1214/07-AOAS148

  3. S. Lundberg, S.-I. Lee, A unified approach to interpreting model predictions (2017). arXiv:1705.07874 [cs.AI].

    Google Scholar 

  4. M.T. Ribeiro, S. Singh, C. Guestrin, “Why should I trust you?”: explaining the predictions of any classifier (2016). arXiv:1602.04938 [cs.LG]

    Google Scholar 

  5. L.S. Shapley, A value for n-person games, in Contributions to the Theory of Games (AM-28), Volume II, ed. by H.W. Kuhn, A.W. Tucker (Princeton University Press, Princeton, 1953). ISBN: 978-1-40088-197-0

    Google Scholar 

  6. D. Slack et al., Fooling LIME and SHAP: adversarial attacks on post hoc explanation methods (2020). arXiv:1911.02508 [cs.LG]

    Google Scholar 

  7. M. Sundararajan, A. Najmi, The many Shapley values for model explanation (2019). arxiv:1908.08474, Comment: 9 pages

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 2021 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Kamath, U., Liu, J. (2021). Post-Hoc Interpretability and Explanations. In: Explainable Artificial Intelligence: An Introduction to Interpretable Machine Learning. Springer, Cham. https://doi.org/10.1007/978-3-030-83356-5_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-83356-5_5

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-83355-8

  • Online ISBN: 978-3-030-83356-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics