Abstract
This chapter presents a method for interpreting neural networks which imposes minimal restrictions on the neural network design. The chapter demonstrates techniques for interpreting a feedforward network, including how to rank the importance of the features. An example demonstrating how to apply interpretability analysis to deep learning models for factor modeling is also presented.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
If Lipschitz continuity is not imposed, then a small change in one of the input values could result in an undesirable large variation in the derivative.
- 2.
When σ is an identity function, the Jacobian J(I (1)) = W (1).
References
Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., et al. (2016). Tensor flow: A system for large-scale machine learning. In Proceedings of the 12th USENIX Conference on Operating Systems Design and Implementation, OSDI’16 (pp. 265–283).
Dimopoulos, Y., Bourret, P., & Lek, S. (1995, Dec). Use of some sensitivity criteria for choosing networks with good generalization ability. Neural Processing Letters,2(6), 1–4.
Dixon, M. F., & Polson, N. G. (2019). Deep fundamental factor models.
Garson, G. D. (1991, April). Interpreting neural-network connection weights. AI Expert,6(4), 46–51.
Greenwell, B. M., Boehmke, B. C., & McCarthy, A. J. (2018, May). A simple and effective model-based variable importance measure. arXiv e-prints, arXiv:1805.04755.
Nielsen, F., & Bender, J. (2010). The fundamentals of fundamental factor models. Technical Report 24, MSCI Barra Research Paper.
Olden, J. D., & Jackson, D. A. (2002). Illuminating the “black box”: a randomization approach for understanding variable contributions in artificial neural networks. Ecological Modelling,154(1), 135–150.
Rosenberg, B., & Marathe, V. (1976). Common factors in security returns: Microeconomic determinants and macroeconomic correlates. Research Program in Finance Working Papers 44, University of California at Berkeley.
Author information
Authors and Affiliations
Appendix
Appendix
Other Interpretability Methods
Partial Dependence Plots (PDPs) evaluate the expected output w.r.t. the marginal density function of each input variable, and allow the importance of the predictors to be ranked. More precisely, partitioning the data X into an interest set, X s, and its complement, \(\boldsymbol {X}_c = \mathcal {X} \setminus \boldsymbol {X}_s\), then the “partial dependence” of the response on X s is defined as
where \(p_{c}\left (\boldsymbol {X}_c\right )\) is the marginal probability density of X c: \(p_{c}\left (\boldsymbol {X}_c\right ) = \int p\left (\boldsymbol {x}\right )d\boldsymbol {x}_s\). Equation (5.20) can be estimated from a set of training data by
where X i,c \(\left (i = 1, 2, \dots , n\right )\) are the observations of X c in the training set; that is, the effects of all the other predictors in the model are averaged out. There are a number of challenges with using PDPs for model interpretability. First, the interaction effects are ignored by the simplest version of this approach. While Greenwell et al. (2018) propose a methodology extension to potentially address the modeling of interactive effects, PDPs do not provide a 1-to-1 correspondence with the coefficients in a linear regression. Instead, we would like to know, under strict control conditions, how the fitted weights and biases of the MLP correspond to the fitted coefficients of linear regression. Moreover in the context of neural networks, by treating the model as a black-box, it is difficult to gain theoretical insight in to how the choice of the network architecture affects its interpretability from a probabilistic perspective.
Garson (1991) partitions hidden-output connection weights into components associated with each input neuron using absolute values of connection weights. Garson’s algorithm uses the absolute values of the connection weights when calculating variable contributions, and therefore does not provide the direction of the relationship between the input and output variables.
Olden and Jackson (2002) determines the relative importance, r ij = [R]ij, of the ith output to the jth predictor variable of the model as a function of the weights, according to the expression
The approach does not account for non-linearity introduced into the activation, which is the most critical aspects of the model. Furthermore, the approach presented was limited to a single hidden layer.
Proof of Variance Bound on Jacobian
Proof
The Jacobian can be written in matrix element form as
where \(c_k:=c_{ijk}:=w^{(2)}_{ik}w^{(1)}_{kj}\) and \(H_k(I):=H(I_k^{(1)})\) is the Heaviside function. As a linear combination of indicator functions, we have
Alternatively, the Jacobian can be expressed in terms of a weighted sum of independent Bernoulli trials involving X:
Without loss of generality, consider the case when p = 1, the dimension of the input space is one. Then Eq. 5.25 simplifies to:
where \(x_k:=-\frac {b^{(1)}_k}{W_k^{(1)}}\). The expectation of the Jacobian is given by
where \(p_k:=\Pr (x_k<X\leq x_{k+1})~\forall k=1,\dots , n-1, ~p_n:=\Pr (x_n<X).\) For finite weights, the expectation is bounded above by \(\sum _{k=1}^{n}a_k\). We can write the variance of the Jacobian as:
Under the assumption that the mean of the Jacobian is invariant to the number of hidden units, or if the weights are constrained so that the mean is constant, then the weights are \(a_k=\frac {\mu _{ij}}{np_k}\). Then the variance is bounded by the mean:
If we relax the assumption that μ ij is independent of n then, under the original weights \(a_k:=\sum _{i=1}^k c_i\):
□
Russell 3000 Factor Model Description
Python Notebooks
The notebooks provided in the accompanying source code repository are designed to gain familiarity with how to implement interpretable deep networks. The examples include toy simulated data and a simple factor model. Further details of the notebooks are included in the README.md file.
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Dixon, M.F., Halperin, I., Bilokon, P. (2020). Interpretability. In: Machine Learning in Finance. Springer, Cham. https://doi.org/10.1007/978-3-030-41068-1_5
Download citation
DOI: https://doi.org/10.1007/978-3-030-41068-1_5
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-41067-4
Online ISBN: 978-3-030-41068-1
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)