Measuring and testing dependence by correlation of distances

  title={Measuring and testing dependence by correlation of distances},
  author={G'abor J. Sz'ekely and Maria L. Rizzo and Nail K. Bakirov},
  journal={Annals of Statistics},
Distance correlation is a new measure of dependence between random vectors. Distance covariance and distance correlation are analogous to product-moment covariance and correlation, but unlike the classical definition of correlation, distance correlation is zero only if the random vectors are independent. The empirical distance dependence measures are based on certain Euclidean distances between sample elements rather than sample moments, yet have a compact representation analogous to the… Expand

Figures and Tables from this paper

Partial Distance Correlation
Partial distance correlation measures association between two random vectors with respect to a third random vector, analogous to, but more general than (linear) partial correlation. DistanceExpand
Conditional Distance Variance and Correlation
Recently a new dependence measure, the distance correlation, has been proposed to measure the dependence between continuous random variables. A nice property of this measure is that it can beExpand
Partial Distance Correlation with Methods for Dissimilarities
Distance covariance and distance correlation are scalar coefficients that characterize independence of random vectors in arbitrary dimension. Properties, extensions, and applications of distanceExpand
On the uniqueness of distance covariance
Distance covariance and distance correlation are non-negative real numbers that characterize the independence of random vectors in arbitrary dimensions. In this work we prove that distance covarianceExpand
Generalizing distance covariance to measure and test multivariate mutual dependence via complete and incomplete V-statistics
This work defines the empirical measures and simplified empirical measures as a trade-off between the complexity and power when testing mutual independence, and captures non-linear and non-monotone mutual dependence between the random vectors. Expand
The distance correlation t-test of independence in high dimension
A modified distance correlation statistic is proposed, such that under independence the distribution of a transformation of the statistic converges to Student t, as dimension tends to infinity, and the resulting t-test is unbiased for every sample size greater than three and all significance levels. Expand
Brownian distance covariance
Distance correlation is a new class of multivariate dependence coefficients applicable to random vectors of arbitrary and not necessarily equal dimension. Distance covariance and distance correlationExpand
On Possibilistic Version of Distance Covariance and Correlation
Distance correlation is a relatively new measure of dependence in probability theory and statistics, which has the great advantage that it gives zero if and only if the variables are independent. InExpand
Distance covariance for stochastic processes
The distance covariance of two random vectors is a measure of their dependence. The empirical distance covariance and correlation can be used as statistical tools for testing whether two randomExpand
Distance Metrics for Measuring Joint Dependence with Application to Causal Inference
Abstract Many statistical applications require the quantification of joint dependence among more than two random vectors. In this work, we generalize the notion of distance covariance to quantifyExpand


A multivariate nonparametric test of independence
A new nonparametric approach to the problem of testing the joint independence of two or more random vectors in arbitrary dimension is developed based on a measure of association determined byExpand
A new test for multivariate normality
We propose a new class of rotation invariant and consistent goodness-of-fit tests for multivariate distributions based on Euclidean distance between sample elements. The proposed test applies to anyExpand
Hierarchical Clustering via Joint Between-Within Distances: Extending Ward's Minimum Variance Method
A hierarchical clustering method that minimizes a joint between-within measure of distance between clusters, by defining a cluster distance and objective function in terms of Euclidean distance, or any power of Euclidesan distance in the interval (0,2). Expand
Correlational Meta-Analysis: Independent and Nonindependent Cases
The purpose of this study was to determine the effect of the violation of the assumption of independence when combining correlation coefficients in a meta-analysis. In this Monte Carlo simulation theExpand
On the Independence of k Sets of Normally Distributed Statistical Variables
IN SUCH fields of investigation as economics, psychology, and anthropology, where observations on several variables are taken into account simultaneously, it is at least as important to studyExpand
This paper attempts to introduce some distribution-free and robust techniques to ecologists and to offer a critical appraisal of the potential advantages and drawbacks of these methods. Expand
Extremal probabilities for Gaussian quadratic forms
Abstract. Denote by Q an arbitrary positive semidefinite quadratic form in centered Gaussian random variables such that E(Q)=1. We prove that for an arbitrary x>0, infQP(Q≤x)=P(χ2n/n≤x), where χn2 isExpand
Group Invariance in Statistical Inference
Group invariance matrices, groups and Jacobians invariance equivariant estimation in curved models some best invarient tests in multinormals minimax tests in multinormals locally minimax tests inExpand
Limitations of the case-only design for identifying gene-environment interactions.
The authors explore the robustness of this procedure to uncertainty about the independence assumption by using simulations and demonstrate that inferences about the multiplicative interaction with the case-only design can be highly distorted when there is departure from the independent assumption. Expand
Multivariate Analysis
An Introduction to Multivariate Statistical AnalysisBy Prof. T. W. Anderson. (Wiley Publications in Mathematical Statistics.) Pp. xii + 374. (New York: John Wiley and Sons, Inc.; London: Chapman andExpand