Skip to main content

Univariate and Multivariate Distributions

  • Chapter
  • First Online:
Risk Measurement

Abstract

In the previous chapters, the notion of risk measure has been introduced and we observe that the key point is the choice of the distributions which characterise the variables. In this chapter several methodologies to fit univariate or multivariate classes of distributions on data sets are introduced. The presentation is not exhaustive. The question of estimation is not documented here and references will be provided “au fil de l’eau”.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
USD 99.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    This property involves that g′(S X(x)) becomes smaller for large values of the random variable X.

  2. 2.

    With abuse of terminology, the term modal is used for local maxima of the Lila density function at which the density’s derivative vanishes and the second order derivative is negative.

  3. 3.

    Sometimes \(F_X^{-1}\) is referred to as quantile function of X.

  4. 4.

    We recall that the above method to construct a new distribution function is related to the method of transformation of random variables, that is using an invertible transformation function y(x) to transform a random variable X into another one Y . Then the density function of the new random variable Y can be implied from the transformation function. Thus, this method directly distorts the distribution of the random variable X.

  5. 5.

    http://www.flong.com/texts/code/shapers_poly/.

  6. 6.

    http://www.flong.com/texts/code/shapers_poly/.

  7. 7.

    A smooth function is a function that has derivatives of all orders everywhere in its domain.

  8. 8.

    Here, the parameters of g(x) are chosen arbitrarily. The purpose is to show the multimodal behaviour of Cobb’s family, and the relationship between the locations of modes and the roots of g(x).

References

  • Cambanis, Stamatis, Steel Huang, and Gordon Simons. 1981. “On the theory of elliptically contoured distributions”. Journal of Multivariate Analysis 11, no. 3: 368–385.

    Article  Google Scholar 

  • Chernobai, A., S. T. Rachev, and F. J. Fabozzi. 2007. Operational risk: A guide to basel II capital requirements, models, and analysis. New York: Wiley.

    Google Scholar 

  • Cherubini, Umberto and Luciano, Elisa and Vecchiato, Walter. 2004. Copula methods in finance. John Wiley & Sons. New York.

    Book  Google Scholar 

  • Clayton, D. 1978. “A model for association in bivariate life tables and its application in epidemiological studies of familial tendency in chronic disease incidence”. Biometrika 65: 141–151.

    Article  Google Scholar 

  • Cobb, Loren, Peter Koppstein, and Neng Hsin Chen. 1983. “Estimation and moment recursion relations for multimodal distributions of the exponential family”. Journal of the American Statistical Association 78, no. 381: 124–130.

    Article  Google Scholar 

  • Deheuvels, P. 1984. “Probabilistic aspects of multivariate extremes”. In Statistical extremes and applications, ed. J. Tiago de Oliveira. Dordrecht: Reidel.

    Google Scholar 

  • Dempster, A. P., N. M. Laird, and D. B. Rubin. 1977. “Maximum likelihood from incomplete data via the EM algorithm”. Journal of the Royal Statistical Society: Series B 39: 1–38.

    Google Scholar 

  • Embrechts, P., C. Klüppelberg, and T. Mikosh. 1997. Modelling extremal events: For insurance and finance. Berlin: Springer.

    Book  Google Scholar 

  • Fang, Hong-Bin, Kai-Tai Fang, and Samuel Kotz. 2002. “The meta-elliptical distributions with given marginals”. Journal of Multivariate Analysis 82, no. 1: 1–16.

    Article  Google Scholar 

  • Fisher, R. A., and L. H. C. Tippett. 1928. “Limiting forms of frequency distributions of the largest or smallest member of a sample”. Mathematical Proceedings of the Cambridge Philosophical Society 24: 180–190.

    Article  Google Scholar 

  • Galambos, J. 1978. The asymptotic theory of extreme order statistics. Wiley series in probability and mathematical statistics. New York: Wiley.

    Google Scholar 

  • Genest, C., K. Ghoudi, and L-P. Rivest. 1995. “A semiparametric estimation procedure of dependence parameters in multivariate families of distributions”. Biometrika 82: 543–552.

    Article  Google Scholar 

  • Gudendorf, G., and J. Segers. 2010. “Extreme-value copulas”. In Proceedings of the workshop on copula theory and its applications held in Warsaw, ed. P. Jaworski. New York: Springer Media.

    Google Scholar 

  • Guégan, Dominique, and Bertrand Hassani. 2015. “Distortion risk measure or the transformation of unimodal distributions into multimodal functions”. In Future perspectives in risk models and finance, 71–88. Cham: Springer.

    Google Scholar 

  • Gumbel, E. 1960. “Distributions des valeurs extrêmes en plusieurs dimensions”. Publications de l’Institut de Statistique de l’Université de Paris 9: 171–173.

    Google Scholar 

  • Hansen, Lars Peter. 1982. “Large sample properties of generalized method of moments estimators”. Econometrica: Journal of the Econometric Society 50: 1029–1054.

    Article  Google Scholar 

  • Hassani, Bertrand K., and Wei Yang. 2016. “The Lila distribution and its applications in risk modelling”. In Documents de travail du Centre d’Economie de la Sorbonne: halshs-01400186, Université Paris 1 Panthéon-Sorbonne, Centre d’Economie de la Sorbonne.

    Google Scholar 

  • Hogg, Robert V., and Allen T. Craig. 1995. Introduction to mathematical statistics, 5th edn. Upper Saddle River: Prentice Hall.

    Google Scholar 

  • Joe, H. 1997. Multivariate models and dependence concepts. Monographs on statistics, applied probability. London: Chapman and Hall.

    Google Scholar 

  • –. 2005. “Asymptotic efficiency of the two-stage estimation method for copula-based models”. Journal of Multivariate Analysis 94: 401–419.

    Google Scholar 

  • Kahneman, Daniel. 1979. “Prospect theory: An analysis of decisions under risk”. Econometrica 47: 278.

    Article  Google Scholar 

  • Kahneman, Daniel, and Amos Tversky. 1979. “On the interpretation of intuitive probability: A reply to Jonathan Cohen”. Cognition 7, no. 4: 409–411.

    Article  Google Scholar 

  • Li, Kehan. 2017. ”Stress, uncertainty and multimodality of risk measures”. PhD Thesis. Université Paris 1 Panthéon-Sorbonne.

    Google Scholar 

  • McLachlan, G., and T. Krishnan. 1997. The EM algorithm and extensions. London: Wiley.

    Google Scholar 

  • Millar, Russell B. 2011. Maximum likelihood estimation and inference: With examples in R, SAS and ADMB, vol. 111. New York: Wiley.

    Book  Google Scholar 

  • Nelsen, R. B. 2006. An introduction to copulas. Springer Series in Statistics, Berlin: Springer.

    Google Scholar 

  • Savu, C., and M. Trede. 2006. “Hierarchical Archimedean copulas”. In Working Paper, University of Münster.

    Google Scholar 

  • Shih, J., and T. Louis. 1995. “Inferences on the association parameter in copula models for bivariate survival data”. Biometrics 51: 1384–1399.

    Article  Google Scholar 

  • Silverman, Bernard W. 2018. Density estimation for statistics and data analysis. Abingdon: Routledge.

    Book  Google Scholar 

  • Sklar, A. 1959. “Fonctions de répartition à n dimensions et leurs marges”. Publications de l’Institut Statistique de l’Université de Paris 8: 229–231.

    Google Scholar 

  • Wang, Shaun S. 2000. “A class of distortion operators for pricing financial and insurance risks”. Journal of Risk and Insurance 67: 15–36.

    Article  Google Scholar 

  • Wang, Shaun S., and Virginia R. Young. 1998. “Risk-adjusted credibility premiums using distorted probabilities”. Scandinavian Actuarial Journal 1998, no. 2: 143–165.

    Article  Google Scholar 

  • Whelan, N. 2004. “Sampling from Archimedean copulas”. Quantitative Finance 4: 339–352.

    Article  Google Scholar 

  • Wirch, Julia L., and Mary R. Hardy. 2001. “Distortion risk measures: Coherence and stochastic dominance”. In International Congress on Insurance: Mathematics and Economics, 15–17.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Appendix: Estimation

Appendix: Estimation

Though estimation procedures are not the focal point of this book, risk measures cannot be assessed in practice without fitting the distributions, either univariate or multivariate. Therefore, we briefly introduce in the following the main approaches to adjust the distributions presented before on data samples.

4.1.1 I Distribution’s Parameters Estimation

  • Maximum likelihood estimation (MLE): the method of maximum likelihood is based on the likelihood function, \(\mathcal {L}(\theta ;x)\). Assuming a distribution f(⋅;θ), where θ ∈ Θ denotes a set of parameters, and Θ represents the domain of possible parameters. The method of maximum likelihood aims at finding the parameters, θ, that maximise the likelihood function, \(\mathcal {L}(\theta ;x)\).

    The method defines a maximum likelihood estimate as follows:

    $$\displaystyle \begin{aligned} \hat{\theta}\in \{{\underset {\theta \in \Theta }{\operatorname {arg\,max} }}\ {\mathcal {L}}(\theta \,;x)\}, \end{aligned} $$
    (1)

    if a maximum exists.

    In practice, it is often convenient (under specific assumptions) to work with the natural logarithm of the likelihood function, called the log-likelihood:

    $$\displaystyle \begin{aligned} \ell (\theta \,;x)=\ln {\mathcal {L}}(\theta \,;x), \end{aligned} $$
    (2)

    As the logarithm function is monotonic and strictly increasing, the MLE is the same regardless of whether the likelihood or the log-likelihood function is maximised. For the properties of the maximum likelihood, we refer to Millar 2011.

  • Generalised method of moments (GMM): Let {Y t}t = 1, …, T, where each observation Y t is an n-dimensional multivariate random variable. Considering a distribution driven by a parameter θ ∈ Θ, the objective is to estimate θ.

    GMM assumes that Y t are generated by a weakly stationary stochastic process (the state independent and identically distributed is a particular case of this condition).

    In order to apply the GMM, the “moment conditions” has to be fulfilled, in other words, the distribution has to have moment. Therefore, the vector-valued function g(Y t, θ) associated to the r.v. has to be known such that

    $$\displaystyle \begin{aligned} m(\theta _{0})\equiv \operatorname{E}[\,g(Y_t,\theta _{0})\,]=0, \end{aligned} $$
    (3)

    where E denotes expectation.

    The objective of the approach is to replace the theoretical expected value with its empirical average given as follows, assuming that we observe Y 1, ..., Y T:

    $$\displaystyle \begin{aligned} {\hat {m}}(\theta )\equiv {\frac {1}{T}}\sum _{{t=1}}^{T}g(Y_{t},\theta ) \end{aligned} $$
    (4)

    and then to minimise this expression with respect to θ.

    Considering the law of large numbers, \({\hat {m}}(\theta )\,\approx \;\operatorname {E}[g(Y_{t},\theta )]\,=\,m(\theta )\) for large values of T, and thus \({\hat {m}}(\theta )\;\approx \;m(\theta _{0})\;=\;0\). The generalised method of moments aims at obtaining \({\hat {\theta }}\) which makes the quantity \({\hat {m}}({\hat \theta })\) as close to zero as possible. Formally, this is equivalent to minimising a particular norm of \({\hat {m}}(\theta )\). The family of norms considered is defined as follows:

    $$\displaystyle \begin{aligned} \|{\hat {m}}(\theta )\|{}_{W}^{2}={\hat {m}}(\theta )^{\mathsf {T}}\,W{\hat {m}}(\theta ), \end{aligned} $$
    (5)

    where W is a positive-definite weighting matrix, and m T its transposition. In practice, the weighting matrix W is obtained using the available data, and therefore is an estimate denoted \(\hat {W}\). The GMM estimator can therefore be formalised as follows:

    $$\displaystyle \begin{aligned} \hat{\theta}={\operatorname{arg}\,min}_{\theta \in \Theta}\left({\frac{1}{T}}\sum_{t=1}^{T}{g(Y_{t},\theta)}\right)^{\mathsf{T}}{\hat{W}}\left({\frac{1}{T}}\sum_{t=1}^{T}{g(Y_{t},\theta)}\right) \end{aligned} $$
    (6)

    Note that under suitable conditions this estimator is consistent, asymptotically normal, and with an appropriate weighting matrix \({\hat {W}}\), asymptotically efficient. For more details on the properties of the GMM, we refer to Hansen 1982.

  • Truncated distributions: Parameter estimation—In practice considering data sets are either incomplete or truncated, therefore the following approaches might be of interest to parametrise the distributions

    • Constrained maximisation likelihood: In the constrained maximum likelihood function approach (Chernobai et al. 2007), one can obtain the parameters of the distribution by directly maximising the likelihood function. If the observed data sample X = X 1, X 2, …, X n constitutes an i.i.d. left-truncated sample, then the likelihood function is expressed as:

      $$\displaystyle \begin{aligned} L(x;\theta|X \leq H) = \prod^{n}_{j=1}{\frac{f(x_{j};\theta|X \neq H)}{P(\theta;X \leq H)}}, \end{aligned} $$
      (7)

      where γ is the parameter set that defines the density.

    • Expectation maximisation algorithm: This algorithm (Dempster et al. 1977; McLachlan and Krishnan 1997) is used to estimate unknown parameters by maximising the likelihood function expectation using available information on truncated data.

      The EM algorithm is an iterative process split up in two steps. In the initial step the set of parameters Θ are initialised at θ 0 values. Likelihood function values are replaced by their expected values.

      In the next step, we maximise this function to obtain new parameters values. These ones are used to re-initialise θ 0 in the first step and then re-iterate the process. The algorithm is:

Algorithm 1

  • Initialisation step: We choose ς 0 , and provide an estimation of the missing data, n missing.

  • Expectation step: Given ς 0 , we compute:

    $$\displaystyle \begin{aligned} Q = E_{\varsigma}\left[\log \frac{L_{\varsigma}(x^{all})}{x^{observed}}\right] = n^{missed} \times E_{\varsigma}[\log f_{\varsigma}(x^{truncated})] + \sum^{n}_{j=1}{\log f_{\varsigma}(x_{j}^{observed})} \end{aligned} $$
    (8)
    • We estimate n missing, \(n^{missing}=n^{observed}\frac {1-P_{\varsigma _{0}}(x < u)}{P_{\varsigma _{0}}(x < u)}\)

    • Random generation of n missing values from the theoretical distribution chosen to fit the head, with respect to the constraint x missing < u

    • x = (x observed, x missing)

    • Log-likelihood function computation

  • Maximisation step: We compute \(\varsigma _{1}=\arg _{\varsigma } \max E_{\varsigma }[\log \frac {L_{\varsigma }(x^{all})}{x^{observed}}]\)

  • If, we have a convergence, we have ς optim . If we do not \(\sqrt {(\varsigma _{1})^{2} - (\varsigma _{0})^{2}} > eps_{value}\) , we repeat step 2 and 3 with ς 0 = ς 1 until convergence.

4.1.2 I Copulas’ Parameters Estimation

In order to estimate the copulas’ parameters, we can carry out one of the following methods(Cherubini et al. (2004)). Regarding chosen copula characteristics, parameters enable tuning the dependence structure.

  1. 1.

    Maximum likelihood:

    The classical maximum likelihood and the two-stage inference function for margins method proposed by Joe (1997, 2005) belong to the first category. The former is based on the maximisation of the full log-likelihood

    $$\displaystyle \begin{aligned} \sum^{n}_{i=1}{\log(c_{\theta})(F_{1,\beta_{1,n}}(X_{i,1}),\ldots,F_{p,\beta_{p,n}}(X_{i,1}))}+\sum^{n}_{i=1}{\sum^{p}_{j=1}{\log f_{j,\beta_{j}}(X_{i,j})}}, \end{aligned} $$
    (9)

    where c θ and \(f_{1;\beta _{1}},\ldots ,f_{p;\beta _{p}}\) are the probability density functions obtained from C θ and \(f_{1;\beta _{1}},\ldots ,f_{p;\beta _{p}}\), respectively when we observe data sets X i,1, ..., X i,n, i = 1, ..., p. With the aim of decreasing the computational burden associated with maximum likelihood estimation, the inference function for margins method first estimates the parameters β 1, …, β p of the marginal cdfs by β 1,n, …, β p,n, and then θ as the maximiser of

    $$\displaystyle \begin{aligned} \sum^{n}_{i=1}{\log(c_{\theta})(F_{1,\beta_{1,n}}(X_{i,1}),\ldots,F_{p,\beta_{p,n}}(X_{i,p}))} \end{aligned} $$
    (10)
  2. 2.

    Pseudo maximum likelihood:

    As it may be argued that the estimation of θ should not be affected by the choice of the marginal cdfs, many authors advocate the use of the maximum pseudo likelihood estimator studied in Genest et al. (1995) and in Shih and Louis (1995) to estimate this one. It consists in maximising the log pseudo-likelihood:

    $$\displaystyle \begin{aligned} \log L(\theta) = \sum^{n}_{i=1}{\log c_{\theta}(\hat{U}_{i})}, \end{aligned} $$
    (11)

    where the \(\hat {U}_{i}=(\hat {U}_{i,1},\ldots ,\hat {U}_{i,p})\) are pseudo-observations computed from the X i = (X i,1, …, X i,p) by \(\hat {U}_{i,j}= \frac {R_{i,j}}{(n+1)}\), with R i,j being the rank of X i,j among X 1,j, …, X n,j.

Method-of-moment approaches are based on the inversion of a consistent estimator of a moment of the copula C θ. The two best-known moments, Spearman’s rho and Kendall’s tau, are, respectively, given by,

  1. 1.

    Spearman’s Rho:

    $$\displaystyle \begin{aligned} \rho_{\theta}=12 \int_{[0,1]^{2}}{C_{\theta}(u_{1},u_{2})d u_{1}d u_{2}}-3 \end{aligned} $$
    (12)
  2. 2.

    Kendall’s Tau:

    $$\displaystyle \begin{aligned} \tau_{\theta}=4 \int_{[0,1]^{2}}{C_{\theta}(u_{1},u_{2})d u_{1}d u_{2}}-1 \end{aligned} $$
    (13)

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Guégan, D., Hassani, B.K. (2019). Univariate and Multivariate Distributions. In: Risk Measurement. Springer, Cham. https://doi.org/10.1007/978-3-030-02680-6_4

Download citation

Publish with us

Policies and ethics