In
statistics
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
and in
probability theory
Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
, distance correlation or distance covariance is a measure of
dependence between two paired
random vector
In probability, and statistics, a multivariate random variable or random vector is a list or vector of mathematical variables each of whose value is unknown, either because the value has not yet occurred or because there is imperfect knowledge ...
s of arbitrary, not necessarily equal,
dimension
In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coo ...
. The population distance correlation coefficient is zero if and only if the random vectors are
independent. Thus, distance correlation measures both linear and nonlinear association between two random variables or random vectors. This is in contrast to
Pearson's correlation, which can only detect linear association between two
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
s.
Distance correlation can be used to perform a
statistical test
A statistical hypothesis test is a method of statistical inference used to decide whether the data provide sufficient evidence to reject a particular hypothesis. A statistical hypothesis test typically involves a calculation of a test statistic. ...
of dependence with a
permutation test. One first computes the distance correlation (involving the re-centering of Euclidean distance matrices) between two random vectors, and then compares this value to the distance correlations of many shuffles of the data.
Background
The classical measure of dependence, the
Pearson correlation coefficient
In statistics, the Pearson correlation coefficient (PCC) is a correlation coefficient that measures linear correlation between two sets of data. It is the ratio between the covariance of two variables and the product of their standard deviatio ...
,
is mainly sensitive to a linear relationship between two variables. Distance correlation was introduced in 2005 by
Gábor J. Székely in several lectures to address this deficiency of Pearson's
correlation
In statistics, correlation or dependence is any statistical relationship, whether causal or not, between two random variables or bivariate data. Although in the broadest sense, "correlation" may indicate any type of association, in statistics ...
, namely that it can easily be zero for dependent variables. Correlation = 0 (uncorrelatedness) does not imply independence while distance correlation = 0 does imply independence. The first results on distance correlation were published in 2007 and 2009. It was proved that distance covariance is the same as the Brownian covariance. These measures are examples of
energy distances.
The distance correlation is derived from a number of other quantities that are used in its specification, specifically: distance variance, distance standard deviation, and distance covariance. These quantities take the same roles as the ordinary
moments with corresponding names in the specification of the
Pearson product-moment correlation coefficient
In statistics, the Pearson correlation coefficient (PCC) is a correlation coefficient that measures linear correlation between two sets of data. It is the ratio between the covariance of two variables and the product of their standard deviation ...
.
Definitions
Distance covariance
Let us start with the definition of the sample distance covariance. Let (''X''
''k'', ''Y''
''k''), ''k'' = 1, 2, ..., ''n'' be a
statistical sample
In this statistics, quality assurance, and survey methodology, sampling is the selection of a subset or a statistical sample (termed sample for short) of individuals from within a statistical population to estimate characteristics of the whole ...
from a pair of real valued or vector valued random variables (''X'', ''Y''). First, compute the ''n'' by ''n''
distance matrices (''a''
''j'', ''k'') and (''b''
''j'', ''k'') containing all pairwise
distances
:
where , , ⋅ , , denotes
Euclidean norm
Euclidean space is the fundamental space of geometry, intended to represent physical space. Originally, in Euclid's ''Elements'', it was the three-dimensional space of Euclidean geometry, but in modern mathematics there are ''Euclidean spaces'' ...
. Then take all doubly centered distances
:
where
is the -th row mean,
is the -th column mean, and
is the
grand mean of the distance matrix of the sample. The notation is similar for the values. (In the matrices of centered distances (''A''
''j'', ''k'') and (''B''
''j'',''k'') all rows and all columns sum to zero.) The squared sample distance covariance (a scalar) is simply the arithmetic average of the products ''A''
''j'', ''k ''''B''
''j'', ''k'':
:
The statistic ''T''
''n'' = ''n'' dCov
2''n''(''X'', ''Y'') determines a consistent multivariate test of independence of random vectors in arbitrary dimensions. For an implementation see ''dcov.test'' function in the ''energy'' package for
R.
The population value of distance covariance can be defined along the same lines. Let ''X'' be a random variable that takes values in a ''p''-dimensional Euclidean space with probability distribution and let ''Y'' be a random variable that takes values in a ''q''-dimensional Euclidean space with probability distribution , and suppose that ''X'' and ''Y'' have finite expectations. Write
:
Finally, define the population value of squared distance covariance of ''X'' and ''Y'' as
:
One can show that this is equivalent to the following definition:
:
where ''E'' denotes expected value, and
and
are independent and identically distributed. The primed random variables
and
denote
independent and identically distributed (iid) copies of the variables
and
and are similarly iid. Distance covariance can be expressed in terms of the classical Pearson's
covariance
In probability theory and statistics, covariance is a measure of the joint variability of two random variables.
The sign of the covariance, therefore, shows the tendency in the linear relationship between the variables. If greater values of one ...
,
cov, as follows:
:
This identity shows that the distance covariance is not the same as the covariance of distances, ). This can be zero even if ''X'' and ''Y'' are not independent.
Alternatively, the distance covariance can be defined as the weighted
''L''2 norm of the distance between the joint
characteristic function of the random variables and the product of their marginal characteristic functions:
[, Theorem 7, (3.7).]
:
where
,
, and
are the
characteristic functions of ''X'', and ''Y'', respectively, ''p'', ''q'' denote the Euclidean dimension of ''X'' and ''Y'', and thus of ''s'' and ''t'', and ''c''
''p'', ''c''
''q'' are constants. The weight function
is chosen to produce a scale equivariant and rotation
invariant measure that doesn't go to zero for dependent variables.
[ One interpretation of the characteristic function definition is that the variables ''eisX'' and ''eitY'' are cyclic representations of ''X'' and ''Y'' with different periods given by ''s'' and ''t'', and the expression in the numerator of the characteristic function definition of distance covariance is simply the classical covariance of ''eisX'' and ''eitY''. The characteristic function definition clearly shows that
dCov2(''X'', ''Y'') = 0 if and only if ''X'' and ''Y'' are independent.
]
Distance variance and distance standard deviation
The ''distance variance'' is a special case of distance covariance when the two variables are identical. The population value of distance variance is the square root of
:
where , , and are independent and identically distributed random variables, denotes the expected value
In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
, and for function , e.g., .
The ''sample distance variance'' is the square root of
:
which is a relative of Corrado Gini's mean difference introduced in 1912 (but Gini did not work with centered distances).
The ''distance standard deviation'' is the square root of the ''distance variance''.
Distance correlation
The ''distance correlation'' of two random variables is obtained by dividing their ''distance covariance'' by the product of their ''distance standard deviations''. The distance correlation is the square root of
:
and the ''sample distance correlation'' is defined by substituting the sample distance covariance and distance variances for the population coefficients above.
For easy computation of sample distance correlation see the ''dcor'' function in the ''energy'' package for R.
Properties
Distance correlation
Distance covariance
This last property is the most important effect of working with centered distances.
The statistic is a biased estimator of . Under independence of X and Y
:
An unbiased estimator
In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In stat ...
of is given by Székely and Rizzo.
Distance variance
Equality holds in (iv) if and only if one of the random variables or is a constant.
Generalization
Distance covariance can be generalized to include powers of Euclidean distance. Define
:
Then for every , and are independent if and only if . It is important to note that this characterization does not hold for exponent ; in this case for bivariate , is a deterministic function of the Pearson correlation. If and are powers of the corresponding distances, , then sample distance covariance can be defined as the nonnegative number for which
:
One can extend to metric-space-valued random variables
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. The term 'random variable' in its mathematical definition refers ...
and : If has law in a metric space with metric , then define