In
statistics
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, Spearman's rank correlation coefficient or Spearman's ''ρ'' is a number ranging from -1 to 1 that indicates how strongly two sets of ranks are correlated. It could be used in a situation where one only has ranked data, such as a tally of gold, silver, and bronze medals. If a statistician wanted to know whether people who are high ranking in sprinting are also high ranking in long-distance running, they would use a Spearman rank correlation coefficient.
The coefficient is named after
Charles Spearman and often denoted by the Greek letter
(rho) or as
. It is a
nonparametric measure of
rank correlation (
statistical dependence between the
ranking
A ranking is a relationship between a set of items, often recorded in a list, such that, for any two items, the first is either "ranked higher than", "ranked lower than", or "ranked equal to" the second. In mathematics, this is known as a weak ...
s of two
variables). It assesses how well the relationship between two variables can be described using a
monotonic function
In mathematics, a monotonic function (or monotone function) is a function between ordered sets that preserves or reverses the given order. This concept first arose in calculus, and was later generalized to the more abstract setting of or ...
.
The Spearman correlation between two variables is equal to the
Pearson correlation between the rank values of those two variables; while Pearson's correlation assesses linear relationships, Spearman's correlation assesses monotonic relationships (whether linear or not). If there are no repeated data values, a perfect Spearman correlation of +1 or −1 occurs when each of the variables is a perfect monotone function of the other.
Intuitively, the Spearman correlation between two variables will be high when observations have a similar (or identical for a correlation of 1)
rank (i.e. relative position label of the observations within the variable: 1st, 2nd, 3rd, etc.) between the two variables, and low when observations have a dissimilar (or fully opposed for a correlation of −1) rank between the two variables.
Spearman's coefficient is appropriate for both
continuous and discrete
ordinal variables. Both Spearman's
and
Kendall's can be formulated as special cases of a more
general correlation coefficient.
Applications
The coefficient can be used to determine how well data fits a model, like when determining the similarity of text documents.
Definition and calculation
The Spearman correlation coefficient is defined as the
Pearson correlation coefficient between the
rank variables.
[
]
For a sample of size
the
pairs of
raw scores
are converted to ranks
and
is computed as
:
where
:
denotes the conventional Pearson product-moment correlation coefficient, Pearson correlation coefficient operator, but applied to the rank variables,
:
is the covariance of the rank variables,
:
and
are the standard deviations of the rank variables.
Only when all
ranks are ''distinct integers'' (no ties), it can be computed using the popular formula
:
where
:
is the difference between the two ranks of each observation,
:
is the number of observations.
Consider a bivariate sample
with corresponding rank pairs
Then the Spearman correlation coefficient of
is
:
where, as usual,
:
:
:
and
:
We shall show that
can be expressed purely in terms of
provided we assume that there be no ties within each sample.
Under this assumption, we have that
can be viewed as random variables
distributed like a uniformly distributed discrete random variable,
on
Hence
and
where
:
:
and thus
:
(These sums can be computed using the formulas for the
triangular numbers and
square pyramidal numbers, or basic
summation results from
umbral calculus.)
Observe now that
:
Putting this all together thus yields
:
Identical values are usually each assigned
fractional ranks equal to the average of their positions in the ascending order of the values, which is equivalent to averaging over all possible permutations.
If ties are present in the data set, the simplified formula above yields incorrect results: Only if in both variables all ranks are distinct, then
(calculated according to biased variance).
The first equation — normalizing by the standard deviation — may be used even when ranks are normalized to
, 1("relative ranks") because it is insensitive both to translation and linear scaling.
The simplified method should also not be used in cases where the data set is truncated; that is, when the Spearman's correlation coefficient is desired for the top ''X'' records (whether by pre-change rank or post-change rank, or both), the user should use the Pearson correlation coefficient formula given above.
[
]
Related quantities
There are several other numerical measures that quantify the extent of
statistical dependence between pairs of observations. The most common of these is the
Pearson product-moment correlation coefficient
In statistics, the Pearson correlation coefficient (PCC) is a correlation coefficient that measures linear correlation between two sets of data. It is the ratio between the covariance of two variables and the product of their standard deviation ...
, which is a similar correlation method to Spearman's rank, that measures the "linear" relationships between the raw numbers rather than between their ranks.
An alternative name for the Spearman
rank correlation is the "grade correlation";
in this, the "rank" of an observation is replaced by the "grade". In continuous distributions, the grade of an observation is, by convention, always one half less than the rank, and hence the grade and rank correlations are the same in this case. More generally, the "grade" of an observation is proportional to an estimate of the fraction of a population less than a given value, with the half-observation adjustment at observed values. Thus this corresponds to one possible treatment of tied ranks. While unusual, the term "grade correlation" is still in use.
Interpretation
The sign of the Spearman correlation indicates the direction of association between ''X'' (the independent variable) and ''Y'' (the dependent variable). If ''Y'' tends to increase when ''X'' increases, the Spearman correlation coefficient is positive. If ''Y'' tends to decrease when ''X'' increases, the Spearman correlation coefficient is negative. A Spearman correlation of zero indicates that there is no tendency for ''Y'' to either increase or decrease when ''X'' increases. The Spearman correlation increases in magnitude as ''X'' and ''Y'' become closer to being perfectly monotonic functions of each other. When ''X'' and ''Y'' are perfectly monotonically related, the Spearman correlation coefficient becomes 1. A perfectly monotonic increasing relationship implies that for any two pairs of data values and , that and always have the same sign. A perfectly monotonic decreasing relationship implies that these differences always have opposite signs.
The Spearman correlation coefficient is often described as being "nonparametric". This can have two meanings. First, a perfect Spearman correlation results when ''X'' and ''Y'' are related by any
monotonic function
In mathematics, a monotonic function (or monotone function) is a function between ordered sets that preserves or reverses the given order. This concept first arose in calculus, and was later generalized to the more abstract setting of or ...
. Contrast this with the Pearson correlation, which only gives a perfect value when ''X'' and ''Y'' are related by a ''linear'' function. The other sense in which the Spearman correlation is nonparametric is that its exact sampling distribution can be obtained without requiring knowledge (i.e., knowing the parameters) of the
joint probability distribution
A joint or articulation (or articular surface) is the connection made between bones, ossicles, or other hard structures in the body which link an animal's skeletal system into a functional whole.Saladin, Ken. Anatomy & Physiology. 7th ed. McGraw- ...
of ''X'' and ''Y''.
Example
In this example, the arbitrary raw data in the table below is used to calculate the correlation between the
IQ of a person with the number of hours spent in front of
TV per week
ictitious values used
Firstly, evaluate
. To do so use the following steps, reflected in the table below.
# Sort the data by the first column (
). Create a new column
and assign it the ranked values 1, 2, 3, ..., ''n''.
# Next, sort the augmented (with
) data by the second column (
). Create a fourth column
and similarly assign it the ranked values 1, 2, 3, ..., ''n''.
# Create a fifth column
to hold the differences between the two rank columns (
and
).
# Create one final column
to hold the value of column
squared.
With
found, add them to find
. The value of ''n'' is 10. These values can now be substituted back into the equation
:
to give
:
which evaluates to with a
''p''-value = 0.627188 (using the
''t''-distribution).
That the value is close to zero shows that the correlation between IQ and hours spent watching TV is very low, although the negative value suggests that the longer the time spent watching television the lower the IQ. In the case of ties in the original values, this formula should not be used; instead, the Pearson correlation coefficient should be calculated on the ranks (where ties are given ranks, as described above).
Confidence intervals
Confidence intervals for Spearman's ''ρ'' can be easily obtained using the Jackknife Euclidean likelihood approach in de Carvalho and Marques (2012). The confidence interval with level
is based on a Wilks' theorem given in the latter paper, and is given by
:
where
is the
quantile of a chi-square distribution with one degree of freedom, and the
are jackknife pseudo-values. This approach is implemented in the R packag
spearmanCI
Determining significance
One approach to test whether an observed value of ''ρ'' is significantly different from zero (''r'' will always maintain ) is to calculate the probability that it would be greater than or equal to the observed ''r'', given the
null hypothesis, by using a
permutation test. An advantage of this approach is that it automatically takes into account the number of tied data values in the sample and the way they are treated in computing the rank correlation.
Another approach parallels the use of the
Fisher transformation in the case of the Pearson product-moment correlation coefficient. That is,
confidence intervals and
hypothesis tests relating to the population value ''ρ'' can be carried out using the Fisher transformation:
:
If ''F''(''r'') is the Fisher transformation of ''r'', the sample Spearman rank correlation coefficient, and ''n'' is the sample size, then
:
is a
''z''-score for ''r'', which approximately follows a standard
normal distribution
In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is
f(x) = \frac ...
under the
null hypothesis of
statistical independence
Independence is a fundamental notion in probability theory, as in statistics and the theory of stochastic processes. Two events are independent, statistically independent, or stochastically independent if, informally speaking, the occurrence of ...
().
One can also test for significance using
:
which is distributed approximately as
Student's ''t''-distribution with degrees of freedom under the
null hypothesis. A justification for this result relies on a permutation argument.
A generalization of the Spearman coefficient is useful in the situation where there are three or more conditions, a number of subjects are all observed in each of them, and it is predicted that the observations will have a particular order. For example, a number of subjects might each be given three trials at the same task, and it is predicted that performance will improve from trial to trial. A test of the significance of the trend between conditions in this situation was developed by E. B. Page and is usually referred to as
Page's trend test for ordered alternatives.
Correspondence analysis based on Spearman's ''ρ''
Classic
correspondence analysis
Correspondence analysis (CA) is a multivariate statistical technique proposed by Herman Otto Hartley (Hirschfeld) and later developed by Jean-Paul Benzécri. It is conceptually similar to principal component analysis, but applies to categorical ...
is a statistical method that gives a score to every value of two nominal variables. In this way the Pearson
correlation coefficient between them is maximized.
There exists an equivalent of this method, called
grade correspondence analysis, which maximizes Spearman's ''ρ'' or
Kendall's τ.
Approximating Spearman's ''ρ'' from a stream
There are two existing approaches to approximating the Spearman's rank correlation coefficient from streaming data.
The first approach
involves coarsening the joint distribution of
. For continuous
values:
cutpoints are selected for
and
respectively, discretizing
these random variables. Default cutpoints are added at
and
. A count matrix of size
, denoted
, is then constructed where