In
probability theory
Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
and
statistics, the covariance function describes how much two
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the p ...
s change together (their ''
covariance
In probability theory and statistics, covariance is a measure of the joint variability of two random variables. If the greater values of one variable mainly correspond with the greater values of the other variable, and the same holds for the le ...
'') with varying spatial or temporal separation. For a
random field or
stochastic process ''Z''(''x'') on a domain ''D'', a covariance function ''C''(''x'', ''y'') gives the covariance of the values of the random field at the two locations ''x'' and ''y'':
:
The same ''C''(''x'', ''y'') is called the
autocovariance
In probability theory and statistics, given a stochastic process, the autocovariance is a function that gives the covariance of the process with itself at pairs of time points. Autocovariance is closely related to the autocorrelation of the proce ...
function in two instances: in
time series
In mathematics, a time series is a series of data points indexed (or listed or graphed) in time order. Most commonly, a time series is a sequence taken at successive equally spaced points in time. Thus it is a sequence of discrete-time data. E ...
(to denote exactly the same concept except that ''x'' and ''y'' refer to locations in time rather than in space), and in multivariate random fields (to refer to the covariance of a variable with itself, as opposed to the
cross covariance
In probability and statistics, given two stochastic processes \left\ and \left\, the cross-covariance is a function that gives the covariance of one process with the other at pairs of time points. With the usual notation \operatorname E for the ...
between two different variables at different locations, Cov(''Z''(''x''
1), ''Y''(''x''
2))).
Admissibility
For locations ''x''
1, ''x''
2, …, ''x''
''N'' ∈ ''D'' the variance of every linear combination
:
can be computed as
:
A function is a valid covariance function if and only if
[{{cite book, title=Statistics for Spatial Data, first=Noel A.C., last=Cressie, year=1993, publisher=Wiley-Interscience] this variance is non-negative for all possible choices of ''N'' and weights ''w''
1, …, ''w''
''N''. A function with this property is called
positive semidefinite.
Simplifications with stationarity
In case of a weakly
stationary
In addition to its common meaning, stationary may have the following specialized scientific meanings:
Mathematics
* Stationary point
* Stationary process
* Stationary state
Meteorology
* A stationary front is a weather front that is not moving ...
random field, where
:
for any lag ''h'', the covariance function can be represented by a one-parameter function
:
which is called a ''covariogram'' and also a ''covariance function''. Implicitly the ''C''(''x''
''i'', ''x''
''j'') can be computed from ''C''
''s''(''h'') by:
:
The
positive definiteness of this single-argument version of the covariance function can be checked by
Bochner's theorem.
Parametric families of covariance functions
For a given
variance
In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of number ...
, a simple stationary parametric covariance function is the "exponential covariance function"
:
where ''V'' is a scaling parameter (correlation length), and ''d'' = ''d''(''x'',''y'') is the distance between two points. Sample paths of a
Gaussian process
In probability theory and statistics, a Gaussian process is a stochastic process (a collection of random variables indexed by time or space), such that every finite collection of those random variables has a multivariate normal distribution, i.e. ...
with the exponential covariance function are not smooth. The "squared exponential" (or "
Gaussian") covariance function:
:
is a stationary covariance function with smooth sample paths.
The
Matérn covariance function and
rational quadratic covariance function are two parametric families of stationary covariance functions. The Matérn family includes the exponential and squared exponential covariance functions as special cases.
See also
*
Autocorrelation function
Autocorrelation, sometimes known as serial correlation in the discrete time case, is the correlation of a signal with a delayed copy of itself as a function of delay. Informally, it is the similarity between observations of a random variable as ...
*
Correlation function
A correlation function is a function that gives the statistical correlation between random variables, contingent on the spatial or temporal distance between those variables. If one considers the correlation function between random variables re ...
*
Covariance matrix
In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements o ...
*
Kriging
In statistics, originally in geostatistics, kriging or Kriging, also known as Gaussian process regression, is a method of interpolation based on Gaussian process governed by prior covariances. Under suitable assumptions of the prior, kriging g ...
*
Positive-definite kernel
*
Random field
*
Stochastic process
*
Variogram
References
Geostatistics
Spatial analysis
Covariance and correlation