Measuring and testing dependence by correlation of distances

  title={Measuring and testing dependence by correlation of distances},
  author={G'abor J. Sz'ekely and Maria L. Rizzo and Nail K. Bakirov},
  journal={Annals of Statistics},
Distance correlation is a new measure of dependence between random vectors. Distance covariance and distance correlation are analogous to product-moment covariance and correlation, but unlike the classical definition of correlation, distance correlation is zero only if the random vectors are independent. The empirical distance dependence measures are based on certain Euclidean distances between sample elements rather than sample moments, yet have a compact representation analogous to the… 

Figures and Tables from this paper

Partial Distance Correlation

Partial distance correlation measures association between two random vectors with respect to a third random vector, analogous to, but more general than (linear) partial correlation. Distance

Conditional Distance Variance and Correlation

Recently a new dependence measure, the distance correlation, has been proposed to measure the dependence between continuous random variables. A nice property of this measure is that it can be

Partial Distance Correlation with Methods for Dissimilarities

This work defines the partial distance correlation statistics with the help of a new Hilbert space where the squared distance covariance is the inner product, and develops and implements a test for zero partialdistance correlation.

The distance correlation t-test of independence in high dimension

Brownian distance covariance

Distance correlation is a new class of multivariate dependence coefficients applicable to random vectors of arbitrary and not necessarily equal dimension. Distance covariance and distance correlation

On Possibilistic Version of Distance Covariance and Correlation

The possibilistic distance correlation is defined, each \(\gamma \)-level set of the joint possibility distribution is equipped with a uniform probability distribution, and the probabilistic distance covariance and correlation between the marginal distributions are determined.

Distance covariance for stochastic processes

An analogs of the distance covariance for two stochastic processes defined on some interval is proposed, which can be used to test the independence of two processes.

Distance Metrics for Measuring Joint Dependence with Application to Causal Inference

The high-order distance covariance is introduced to measure the so-called Lancaster interaction dependence and is defined as a linear combination of pairwise distance covariances and their higher-order counterparts which together completely characterize mutual independence.

Measuring Linear Correlation Between Random Vectors

  • Giovanni Puccetti
  • Mathematics
    Information Sciences
  • 2022
We introduce a new scalar coefficient to measure linear correlation between random vectors which preserves all the relevant properties of Pearson’s correlation in arbitrary dimensions. The new



Hierarchical Clustering via Joint Between-Within Distances: Extending Ward's Minimum Variance Method

A hierarchical clustering method that minimizes a joint between-within measure of distance between clusters, by defining a cluster distance and objective function in terms of Euclidean distance, or any power of Euclidesan distance in the interval (0,2).

Correlational Meta-Analysis: Independent and Nonindependent Cases

The purpose of this study was to determine the effect of the violation of the assumption of independence when combining correlation coefficients in a meta-analysis. In this Monte Carlo simulation the

On the Independence of k Sets of Normally Distributed Statistical Variables

IN SUCH fields of investigation as economics, psychology, and anthropology, where observations on several variables are taken into account simultaneously, it is at least as important to study


This paper attempts to introduce some distribution-free and robust techniques to ecologists and to offer a critical appraisal of the potential advantages and drawbacks of these methods.

Extremal probabilities for Gaussian quadratic forms

Abstract. Denote by Q an arbitrary positive semidefinite quadratic form in centered Gaussian random variables such that E(Q)=1. We prove that for an arbitrary x>0, infQP(Q≤x)=P(χ2n/n≤x), where χn2 is

Group Invariance in Statistical Inference

Group invariance matrices, groups and Jacobians invariance equivariant estimation in curved models some best invarient tests in multinormals minimax tests in multinormals locally minimax tests in

Limitations of the case-only design for identifying gene-environment interactions.

The authors explore the robustness of this procedure to uncertainty about the independence assumption by using simulations and demonstrate that inferences about the multiplicative interaction with the case-only design can be highly distorted when there is departure from the independent assumption.

Multivariate Analysis

An Introduction to Multivariate Statistical AnalysisBy Prof. T. W. Anderson. (Wiley Publications in Mathematical Statistics.) Pp. xii + 374. (New York: John Wiley and Sons, Inc.; London: Chapman and

Gaussian Measures in Banach Spaces

Gaussian measures in Banach spaces.- Equivalence and orthogonality of Gaussian measures.- Some results about abstract Wiener space.