Isserlis Gaussian Moment Theorem
   HOME

TheInfoList



OR:

In
probability theory Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
, Isserlis's theorem or Wick's probability theorem is a formula that allows one to compute higher-order moments of the
multivariate normal distribution In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional ( univariate) normal distribution to higher dimensions. One d ...
in terms of its covariance matrix. It is named after
Leon Isserlis Leon Isserlis (1881–1966) was a Russian-born British statistician known for his work on the exact distribution of sample moments, including Isserlis’ theorem. He also brought to the attention of British statisticians the work of Russ ...
. This theorem is also particularly important in
particle physics Particle physics or high-energy physics is the study of Elementary particle, fundamental particles and fundamental interaction, forces that constitute matter and radiation. The field also studies combinations of elementary particles up to the s ...
, where it is known as
Wick's theorem Wick's theorem is a method of reducing high- order derivatives to a combinatorics problem. It is named after Italian physicist Gian Carlo Wick. It is used extensively in quantum field theory to reduce arbitrary products of creation and annihil ...
after the work of . Other applications include the analysis of portfolio returns, quantum field theory and generation of colored noise.


Statement

If (X_1,\dots, X_) is a zero-mean
multivariate normal In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One de ...
random vector, then\operatorname ,X_1 X_2\cdots X_\,= \sum_\prod_ \operatorname ,X_i X_j\,= \sum_\prod_ \operatorname(\,X_i, X_j\,), where the sum is over all the pairings of \, i.e. all distinct ways of partitioning \ into pairs \, and the product is over the pairs contained in p. More generally, if (Z_1,\dots, Z_) is a zero-mean ''complex''-valued multivariate normal random vector, then the formula still holds. The expression on the right-hand side is also known as the
hafnian In mathematics, the hafnian is a scalar function of a symmetric matrix that generalizes the permanent. The hafnian was named by Eduardo R. Caianiello "to mark the fruitful period of stay in Copenhagen (Hafnia in Latin)." Definition The hafni ...
of the covariance matrix of (X_1,\dots, X_).


Odd case

If n=2m+1 is odd, there does not exist any pairing of \. Under this hypothesis, Isserlis's theorem implies that\operatorname ,X_1 X_2\cdots X_\,= 0. This also follows from the fact that -X=(-X_1,\dots,-X_n) has the same distribution as X, which implies that \operatorname ,X_1 \cdots X_\,\operatorname ,(-X_1) \cdots (-X_)\,-\operatorname ,X_1 \cdots X_\,= 0.


Even case

In his original paper,
Leon Isserlis Leon Isserlis (1881–1966) was a Russian-born British statistician known for his work on the exact distribution of sample moments, including Isserlis’ theorem. He also brought to the attention of British statisticians the work of Russ ...
proves this theorem by mathematical induction, generalizing the formula for the 4^ order moments, which takes the appearance : \operatorname ,X_1 X_2 X_3 X_4\,= \operatorname
_1X_2 1X or 1-X may refer to: * 1X Band, a musical group from Slovenia * 1. X. 1905, a piano composition by Leoš Janáček * Saab 9-1X * Alberta Highway 1X; see Alberta Highway 1A * NY 1X; see Hutchinson River Parkway * SSH 1X (WA); see List of former ...
,\operatorname _3X_4+ \operatorname _1X_3,\operatorname _2X_4+ \operatorname
_1X_4 1X or 1-X may refer to: * 1X Band, a musical group from Slovenia * 1. X. 1905, a piano composition by Leoš Janáček * Saab 9-1X * Alberta Highway 1X; see Alberta Highway 1A * NY 1X; see Hutchinson River Parkway * SSH 1X (WA); see List of former ...
,\operatorname
_2X_3 X, or x, is the twenty-fourth letter of the Latin alphabet, used in the modern English alphabet, the alphabets of other western European languages and others worldwide. Its name in English is ''ex'' (pronounced ), plural ''exes''."X", '' ...
If n=2m is even, there exist (2m)!/(2^m!) = (2m-1)!! (see
double factorial In mathematics, the double factorial of a number , denoted by , is the product of all the positive integers up to that have the same Parity (mathematics), parity (odd or even) as . That is, n!! = \prod_^ (n-2k) = n (n-2) (n-4) \cdots. Restated ...
) pair partitions of \: this yields (2m)!/(2^m!) = (2m-1)!! terms in the sum. For example, for 4^ order moments (i.e. 4 random variables) there are three terms. For 6^-order moments there are 3\times 5=15 terms, and for 8^-order moments there are 3\times5\times7 = 105 terms.


Example

We can evaluate the
characteristic function In mathematics, the term "characteristic function" can refer to any of several distinct concepts: * The indicator function of a subset, that is the function \mathbf_A\colon X \to \, which for a given subset ''A'' of ''X'', has value 1 at points ...
of gaussians by the Isserlis theorem:E ^= \sum_k \frac E ^k= \sum_k \frac E ^= \sum_k \frac \frac E ^k = e^


Proof

Since both sides of the formula are multilinear in X_1, ..., X_n, if we can prove the real case, we get the complex case for free. Let \Sigma_ = \operatorname(X_i, X_j) be the covariance matrix, so that we have the zero-mean
multivariate normal In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One de ...
random vector (X_1, ..., X_n) \sim N(0, \Sigma). Since both sides of the formula are continuous with respect to \Sigma, it suffices to prove the case when \Sigma is invertible. Using quadratic factorization -x^T\Sigma^x/2 + v^Tx - v^T\Sigma v/2 = -(x-\Sigma v)^T\Sigma^(x-\Sigma v)/2, we get \frac\int e^ dx = e^ Differentiate under the integral sign with \partial_, _ to obtain E _1\cdots X_n= \partial_, _e^. That is, we need only find the coefficient of term v_1\cdots v_n in the Taylor expansion of e^. If n is odd, this is zero. So let n = 2m, then we need only find the coefficient of term v_1\cdots v_n in the polynomial \frac(v^T\Sigma v/2)^m. Expand the polynomial and count, we obtain the formula. \square


Generalizations


Gaussian integration by parts

An equivalent formulation of the Wick's probability formula is the Gaussian
integration by parts In calculus, and more generally in mathematical analysis, integration by parts or partial integration is a process that finds the integral of a product of functions in terms of the integral of the product of their derivative and antiderivati ...
. If (X_1,\dots X_) is a zero-mean
multivariate normal In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One de ...
random vector, then \operatorname(X_1 f(X_1,\ldots,X_n))=\sum_^ \operatorname(X_1, X_i)\operatorname(\partial_f(X_1,\ldots,X_n)).This is a generalization of
Stein's lemma Stein's lemma, named in honor of Charles Stein, is a theorem of probability theory that is of interest primarily because of its applications to statistical inference — in particular, to James–Stein estimation and empirical Bayes methods & ...
. The Wick's probability formula can be recovered by induction, considering the function f:\mathbb^n\to\mathbb defined by f(x_1,\ldots,x_n)=x_2\ldots x_n. Among other things, this formulation is important in Liouville conformal field theory to obtain
conformal Ward identities A two-dimensional conformal field theory is a quantum field theory on a Euclidean two-dimensional space, that is invariant under local conformal transformations. In contrast to other types of conformal field theories, two-dimensional conformal fie ...
,
BPZ equations A two-dimensional conformal field theory is a quantum field theory on a Euclidean two-dimensional space, that is invariant under local conformal transformations. In contrast to other types of conformal field theories, two-dimensional conformal fie ...
and to prove the Fyodorov-Bouchaud formula.


Non-Gaussian random variables

For non-Gaussian random variables, the moment-
cumulant In probability theory and statistics, the cumulants of a probability distribution are a set of quantities that provide an alternative to the '' moments'' of the distribution. Any two probability distributions whose moments are identical will have ...
s formula replaces the Wick's probability formula. If (X_1,\dots X_) is a vector of
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
s, then \operatorname(X_1 \ldots X_n)=\sum_ \prod_ \kappa\big((X_i)_\big),where the sum is over all the partitions of \, the product is over the blocks of p and \kappa\big((X_i)_\big) is the joint cumulant of (X_i)_.


Uniform distribution on the unit sphere

Consider X = (X_1,\dots,X_d) uniformly distributed on the unit sphere S^, so that \, X\, =1 almost surely. In this setting, the following holds. If n is odd, \operatorname\bigl _\,X_\,\cdots\,X_\bigr\;=\; 0.\! If n = 2k is even, \operatorname\bigl _\,\cdots\,X_\bigr\;=\; \frac \sum_ \prod_ \delta_, where P_^2 is the set of all pairings of \, \delta_ is the
Kronecker delta In mathematics, the Kronecker delta (named after Leopold Kronecker) is a function of two variables, usually just non-negative integers. The function is 1 if the variables are equal, and 0 otherwise: \delta_ = \begin 0 &\text i \neq j, \\ 1 &\ ...
. Since there are , P_^2, =(2k - 1)!! delta-terms, we get on the diagonal: \operatorname ,X_1^\,\;=\; \frac. Here, (2k - 1)!! denotes the
double factorial In mathematics, the double factorial of a number , denoted by , is the product of all the positive integers up to that have the same Parity (mathematics), parity (odd or even) as . That is, n!! = \prod_^ (n-2k) = n (n-2) (n-4) \cdots. Restated ...
. These results are discussed in the context of random vectors and irreducible representations in the work by Kushkuley (2021).


See also

*
Wick's theorem Wick's theorem is a method of reducing high- order derivatives to a combinatorics problem. It is named after Italian physicist Gian Carlo Wick. It is used extensively in quantum field theory to reduce arbitrary products of creation and annihil ...
*
Cumulant In probability theory and statistics, the cumulants of a probability distribution are a set of quantities that provide an alternative to the '' moments'' of the distribution. Any two probability distributions whose moments are identical will have ...
s *
Normal distribution In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is f(x) = \frac ...


References


Further reading

* {{DEFAULTSORT:Isserlis's Theorem Moments (mathematics) Normal distribution Theorems in probability theory