Abstract
Bergsma (A new correlation coefficient, its orthogonal decomposition and associated tests of independence, arXiv preprint arXiv:math/0604627, 2006) proposed a covariance \(\kappa (X,Y)\) between random variables X and Y, and gave two estimates for it, based on n i.i.d. samples. He derived the asymptotic distributions of these estimates under the assumption of independence between X and Y. Our main focus is on the dependent case. This measure turns out to be same as the distance covariance (dCov) measure for multivariate X and Y, when we specialize to real-valued X and Y. We first derive several alternate expressions for \(\kappa\), which are useful to understand the properties of \(\kappa\) and its estimates better. One of the alternate expressions for \(\kappa\) leads to a very intuitive third estimator of \(\kappa\) that is a nice function of four U-statistics. We establish the exact finite sample algebraic relation between the three estimates. This yields the relation between the bias of these estimators. In the dependent case, using the U statistics central limit theorem, it is easy to show that our estimate is asymptotic normal. The relation between the three estimates is then used to show that Bergsma’s two estimates have the same limit distribution in the dependent case. When X and Y are independent, the above limit is degenerate. With a higher scaling, the non-degenerate limit distribution of all three estimators is obtained using the theory of degenerate U-statistics and the above algebraic relations. In particular, the known asymptotic distribution results for the two estimates of Bergsma for the independent case follow. For specific parametric bivariate distributions, the value of \(\kappa\) can be derived in terms of the natural dependence parameters of these distributions. In particular, we derive the formula for \(\kappa\) when (X, Y) are distributed as Gumbel’s bivariate exponential. We bring out various aspects of these estimators through extensive simulations from several prominent bivariate distributions. In particular, we investigate the empirical relationship between \(\kappa\) and the dependence parameters, the distributional properties of the estimators, and the accuracy of these estimators. We also investigate the finite sample powers of these measures for testing independence, compare these among themselves, and with other well known such measures. Based on these exercises, the proposed estimator seems as good or better than its competitors both in terms of power and computing efficiency.
Similar content being viewed by others
References
Bergsma, W. P. (2006). A new correlation coefficient, its orthogonal decomposition and associated tests of independence (p. 46). arXiv preprint arXiv:math/0604627. Accessed 15 Aug 2023.
Chatterjee, S. (2021). A new coefficient of correlation. Journal of the American Statistical Association, 116(536), 2009–2022.
Gretton, A., Bousquet, O., Smola, A., Schölkopf, B. (2005a). Measuring statistical dependence with hilbert-schmidt norms. In: S. Jain, H. U. Simon, E. Tomita (eds). Proceedings of the International Conference on Algorithmic Learning Theory, (pp. 63–77). Springer-Verlag.
Gretton, A., Smola, A., Schölkopf, B., Bousquet, O., Herbrich, R., Belitski, A., Augath, M., Murayama, Y., Pauls, J., & Logothetis, N. (2005b). Kernel constrained covariance for dependence measurement. AISTATS, 10, 112–119.
Gumbel, E. J. (1960). Bivariate exponential distributions. Journal of the American Statistical Association, 55(292), 698–707.
Hoeffding, W. (1948). A non-parametric test of independence. The Annals of Mathematical Statistics, 19(4), 546–557.
Kappara, D., Bose, A., Bhattacharjee, M. (2022). Assessing bivariate independence: revisiting Bergsma’s covariance (p. 33). arXiv preprint arXiv:Math/2212.08921. Accessed 15 Aug 2023.
Kappara, D., Bose, A., Bhattacharjee, M. (2023). Measuring spatial association and testing spatial independence based on short time course data (p. 19). arXiv preprint arXiv:Math/2303.16824. Accessed 15 Aug 2023.
Lee, A. J. (2019). U-statistics: Theory and Practice (2nd ed.). Routledge.
Ma, J. (2022). Evaluating independence and conditional independence measures (p. 54). arXiv:2205.07253v1. Accessed 15 Aug 2023.
Song, L., Smola, A., Gretton, A., Bedo, J., & Borgwardt, K. (2012). Feature selection via dependence maximization. Journal of Machine Learning Research, 13(47), 1393–1434.
Székely, G. J., Rizzo, M. L., & Bakirov, N. K. (2007). Measuring and testing dependence by correlation of distances. The Annals of Statistics, 35(6), 2769–2794.
Tjøstheim, D., Otneim, H., & Stove, B. (2022). Statistical dependence: Beyond Pearson’s \(\rho\). Statistical Science, 37(1), 90–109.
Acknowledgements
The research of Arup Bose has been supported by a J.C. Bose Fellowship from Science Education and Research Board, Government of India. We are grateful to the two Reviewers whose comments have lead to a significant improvement of the manuscript.
Funding
The research of Arup Bose has been supported by a J.C. Bose Fellowship from Science Education and Research Board, Government of India.
Author information
Authors and Affiliations
Contributions
All authors contributed to the study conception, design, material preparation, data collection and analysis.
Corresponding author
Ethics declarations
Competing interest
There are no known competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
1.1 A. Six bivariate distributions
Standard univariate distributions can be extended to multivariate distributions in multiple ways. For illustration we have chosen six distributions. We present the specific bivariate extensions used in our simulations.
Normal: This is simply the bivariate normal distribution with zero means, unit variances and correlation denoted by \(\theta\).
Uniform: A pair of correlated uniform random variables (X, Y) is generated by starting with two independent uniform variables U and V. Consider a dependency parameter \(\theta \in [-1,\ 1]\). Note that \(\theta = \pm 1\) indicate perfect correlation. Thus in that case the uniform pair (X, Y) equals \((U, 1-U)\) or (U, U). In other cases we draw another random number, say W, from Beta\((\alpha , 1)\), where the shape parameter \(\alpha\) equals
Then we define the dependent pair (X, Y) by \(X=U\) and
Exponential: The dependent exponential pair (X, Y) is obtained as follows. Let X be Exponential(1) random variable. Let U be a uniform(0,1) random variable. Then define Y as
where \(E = ((1 - \theta + \theta X)/\exp (X))/(1 + \theta X)\) and \(G = ((\theta + \theta ^2 X)/\exp (X))/((1 + \theta X)^2).\)
Laplace: \(Z = (X, Y)\) is said to be bivariate standard Laplace, if their joint density is as follows.
where \(\Sigma\) is the \(2\times 2\) matrix with diagonal and off-diagonal elements as 1 and \(\theta\) respectively.
Logistic: Let U, V be two correlated uniform random variables as defined above. Define \(X = \log (U) - \log (1-U)\) and \(Y = \log (V) - \log (1-V)\). Then (X, Y) has bivariate logistic distribution.
Chi-square: Let U, V be bivariate mean zero normal rvs with unit variance and correlation \(\theta\). Define \(X = U^{2}\) and \(Y = V^{2}\). Clearly, each has \(\chi ^{2}\) distribution and have correlation \(\theta ^2\) and (X, Y) has bivariate Chi-square distribution.
1.2 B. Discrete approximation
Recall that when X and Y are independent, the asymptotic limits feature eigenvalues that depend on the distribution of X and Y. The eigensystem of the kernel \(h_F\) is the solution to the integral equation,
In general, this equation does not admit a closed form solution. For the case of discrete and continuous F, this is reduced to a simpler problem as follows.
The eigensystem in the discrete case: Let X be a discrete random variable taking values \(x_1< x_2< \cdots < x_{t}\) with \(P(X=x_m)=p_m, 1\le m \le t\). Then we have the following lemma.
Lemma 3
(Bergsma (2006)) The non-zero eigenvalues and eigenvectors of \(h_F\) are the solutions to the equations;
where \(c_m=(x_m-x_{m-1})^{-1}\).
In matrix notation we must solve the generalized eigenvalue problem \(D_pg=\lambda C g\), where \(D_p\) is a diagonal matrix with \(\{p_m\}\) on the main diagonal, g is the eigenvector with corresponding eigenvalue \(\lambda\), and C is a matrix of coefficients \(c_m\) from the above system, i.e.,
Incidentally, the equation given in Bergsma (2006) appears to have some typographical errors in the signs of the entries of his C matrix and the corresponding difference equations. We have made the required corrections in the difference equations and the C matrix above.
The eigensystem in the continuous case:
Lemma 4
(Bergsma (2006)) Suppose F is strictly increasing on the support of the probability distribution and f is the derivative of F. Let g be the eigenfunction corresponding to the integral (7.1). Let \(y(x)=g(F^{-1}(x))\) and suppose y is twice differentiable. Then any eigenvalue \(\lambda\) and its corresponding eigenvector \(y_k(x)\) are solutions to the equation
subject to the condition
For most distributions, Eq. (7.3) does not have a closed from solution. In such cases we take a cue from the disscrete case and use a discrete approximation as follows. Let X be a continuous random variable with distribution function F. For a (large) positive integer t, define a discrete approximation \(X^{(t)}\) to X as
Let \(F^{(t)}\) be the distribution function of \(X^{(t)}\). Then the eigen pair for the kernel \(h_{F^{(t)}}\), obtained by solving (7.2) with the coefficients \(c_m={({x_m}^{(t)}-{x_{m-1}}^{(t)})}^{-1}\), serves as an approximate eigen pair for \(h_F\).
1.3 C. Empirical distributions of the \(\kappa\) estimators
Figure 11 presents the box-plots of the empirical asymptotic distributions of the three estimates with various underlying distributions having varying degrees of dependence.
1.4 D. Empirical power
We now present power computation for tests of independence, which are based on the three estimators of \(\kappa\), as well as on other known measures of dependence.
1.5 E. Computation time
We now present efficiency data for computation of these measures of dependence mentioned earlier.
1.6 F. Code for the \(\kappa\) estimates
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Bose, A., Kappara, D. & Bhattacharjee, M. Estimation of Bergsma’s covariance. J. Korean Stat. Soc. 52, 1025–1054 (2023). https://doi.org/10.1007/s42952-023-00236-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s42952-023-00236-1
Keywords
- Bergsma’s covariance
- Eigenvalues
- U and V-statistic
- Degenerate U-statistics
- Measures of dependence
- Distance covariance
- Powers of tests of independence