HOME

TheInfoList



OR:

In
probability theory Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
and
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, the cumulants of a
probability distribution In probability theory and statistics, a probability distribution is a Function (mathematics), function that gives the probabilities of occurrence of possible events for an Experiment (probability theory), experiment. It is a mathematical descri ...
are a set of quantities that provide an alternative to the '' moments'' of the distribution. Any two probability distributions whose moments are identical will have identical cumulants as well, and vice versa. The first cumulant is the
mean A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
, the second cumulant is the
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
, and the third cumulant is the same as the third central moment. But fourth and higher-order cumulants are not equal to central moments. In some cases theoretical treatments of problems in terms of cumulants are simpler than those using moments. In particular, when two or more random variables are
statistically independent Independence is a fundamental notion in probability theory, as in statistics and the theory of stochastic processes. Two event (probability theory), events are independent, statistically independent, or stochastically independent if, informally s ...
, the th-order cumulant of their sum is equal to the sum of their th-order cumulants. As well, the third and higher-order cumulants of a
normal distribution In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is f(x) = \frac ...
are zero, and it is the only distribution with this property. Just as for moments, where ''joint moments'' are used for collections of random variables, it is possible to define ''joint cumulants''.


Definition

The cumulants of a random variable are defined using the cumulant-generating function , which is the
natural logarithm The natural logarithm of a number is its logarithm to the base of a logarithm, base of the e (mathematical constant), mathematical constant , which is an Irrational number, irrational and Transcendental number, transcendental number approxima ...
of the moment-generating function: K(t)=\log\operatorname\left ^\right The cumulants are obtained from a
power series In mathematics, a power series (in one variable) is an infinite series of the form \sum_^\infty a_n \left(x - c\right)^n = a_0 + a_1 (x - c) + a_2 (x - c)^2 + \dots where ''a_n'' represents the coefficient of the ''n''th term and ''c'' is a co ...
expansion of the cumulant generating function: K(t)=\sum_^\infty \kappa_ \frac =\kappa_1 \frac + \kappa_2 \frac+ \kappa_3 \frac+ \cdots = \mu t + \sigma^2 \frac + \cdots. This expansion is a Maclaurin series, so the th cumulant can be obtained by differentiating the above expansion times and evaluating the result at zero: \kappa_ = K^(0). If the moment-generating function does not exist, the cumulants can be defined in terms of the relationship between cumulants and moments discussed later.


Alternative definition of the cumulant generating function

Some writers prefer to define the cumulant-generating function as the natural logarithm of the
characteristic function In mathematics, the term "characteristic function" can refer to any of several distinct concepts: * The indicator function of a subset, that is the function \mathbf_A\colon X \to \, which for a given subset ''A'' of ''X'', has value 1 at points ...
, which is sometimes also called the ''second'' characteristic function, H(t)=\log\operatorname \left ^\right\sum_^\infty \kappa_n \frac=\mu it - \sigma^2 \frac + \cdots An advantage of — in some sense the function evaluated for purely imaginary arguments — is that is well defined for all real values of even when is not well defined for all real values of , such as can occur when there is "too much" probability that has a large magnitude. Although the function will be well defined, it will nonetheless mimic in terms of the length of its Maclaurin series, which may not extend beyond (or, rarely, even to) linear order in the argument , and in particular the number of cumulants that are well defined will not change. Nevertheless, even when does not have a long Maclaurin series, it can be used directly in analyzing and, particularly, adding random variables. Both the
Cauchy distribution The Cauchy distribution, named after Augustin-Louis Cauchy, is a continuous probability distribution. It is also known, especially among physicists, as the Lorentz distribution (after Hendrik Lorentz), Cauchy–Lorentz distribution, Lorentz(ian) ...
(also called the Lorentzian) and more generally,
stable distribution In probability theory, a distribution is said to be stable if a linear combination of two independent random variables with this distribution has the same distribution, up to location and scale parameters. A random variable is said to be st ...
s (related to the Lévy distribution) are examples of distributions for which the power-series expansions of the generating functions have only finitely many well-defined terms.


Some basic properties

The nth cumulant \kappa_n(X) of (the distribution of) a random variable X enjoys the following properties: * If n>1 and c is constant (i.e. not random) then \kappa_n(X+c) = \kappa_n(X), i.e. the cumulant is translation invariant. (If n=1 then we have \kappa_1(X+c) = \kappa_1(X)+c.) * If c is constant (i.e. not random) then \kappa_n(cX) = c^n\kappa_n(X), i.e. the nth cumulant is
homogeneous Homogeneity and heterogeneity are concepts relating to the uniformity of a substance, process or image. A homogeneous feature is uniform in composition or character (i.e., color, shape, size, weight, height, distribution, texture, language, i ...
of degree n. * If random variables X_1,\ldots,X_m are independent then \kappa_n(X_1+\cdots+X_m) = \kappa_n(X_1) + \cdots + \kappa_n(X_m)\,. That is, the cumulant is cumulative — hence the name. The cumulative property follows quickly by considering the cumulant-generating function: \begin K_(t) & =\log\operatorname \left ^\right\\ pt& = \log \left(\operatorname \left ^\right\cdots \operatorname \left e^ \right\right) \\ pt& = \log\operatorname\left ^\right+ \cdots + \log \operatorname \left e^ \right\\ pt&= K_(t) + \cdots + K_(t), \end so that each cumulant of a sum of independent random variables is the sum of the corresponding cumulants of the addends. That is, when the addends are statistically independent, the mean of the sum is the sum of the means, the variance of the sum is the sum of the variances, the third cumulant (which happens to be the third central moment) of the sum is the sum of the third cumulants, and so on for each order of cumulant. A distribution with given cumulants can be approximated through an Edgeworth series.


First several cumulants as functions of the moments

All of the higher cumulants are polynomial functions of the central moments, with integer coefficients, but only in degrees 2 and 3 are the cumulants actually central moments. * \kappa_1(X) = \operatorname E = mean * \kappa_2(X) = \operatorname(X) = \operatorname E\left 2\right=the variance, or second central moment. * \kappa_3(X) = \operatorname E\left 3\right= the third central moment. * \kappa_4(X) = \operatorname E\left 4\right- 3\left( \operatorname E\left 2\right\right)^2= the fourth central moment minus three times the square of the second central moment. Thus this is the first case in which cumulants are not simply moments or central moments. The central moments of degree more than 3 lack the cumulative property. * \kappa_5(X) = \operatorname E\left 5\right- 10 \operatorname E\left ^3\right\operatorname E\left 2\right


Cumulants of some discrete probability distributions

* The constant random variables . The cumulant generating function is . The first cumulant is and the other cumulants are zero, . * The
Bernoulli distribution In probability theory and statistics, the Bernoulli distribution, named after Swiss mathematician Jacob Bernoulli, is the discrete probability distribution of a random variable which takes the value 1 with probability p and the value 0 with pro ...
s, (number of successes in one trial with probability of success). The cumulant generating function is . The first cumulants are and . The cumulants satisfy a recursion formula \kappa_=p (1-p) \frac. * The
geometric distribution In probability theory and statistics, the geometric distribution is either one of two discrete probability distributions: * The probability distribution of the number X of Bernoulli trials needed to get one success, supported on \mathbb = \; * T ...
s, (number of failures before one success with probability of success on each trial). The cumulant generating function is . The first cumulants are , and . Substituting gives and . * The
Poisson distribution In probability theory and statistics, the Poisson distribution () is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time if these events occur with a known const ...
s. The cumulant generating function is . All cumulants are equal to the parameter: . * The
binomial distribution In probability theory and statistics, the binomial distribution with parameters and is the discrete probability distribution of the number of successes in a sequence of statistical independence, independent experiment (probability theory) ...
s, (number of successes in
independent Independent or Independents may refer to: Arts, entertainment, and media Artist groups * Independents (artist group), a group of modernist painters based in Pennsylvania, United States * Independentes (English: Independents), a Portuguese artist ...
trials with probability of success on each trial). The special case is a Bernoulli distribution. Every cumulant is just times the corresponding cumulant of the corresponding Bernoulli distribution. The cumulant generating function is . The first cumulants are and . Substituting gives and . The limiting case is a Poisson distribution. * The
negative binomial distribution In probability theory and statistics, the negative binomial distribution, also called a Pascal distribution, is a discrete probability distribution that models the number of failures in a sequence of independent and identically distributed Berno ...
s, (number of failures before successes with probability of success on each trial). The special case is a geometric distribution. Every cumulant is just times the corresponding cumulant of the corresponding geometric distribution. The derivative of the cumulant generating function is . The first cumulants are , and . Substituting gives and . Comparing these formulas to those of the binomial distributions explains the name 'negative binomial distribution'. The limiting case is a Poisson distribution. Introducing the variance-to-mean ratio \varepsilon=\mu^\sigma^2=\kappa_1^\kappa_2, the above probability distributions get a unified formula for the derivative of the cumulant generating function: K'(t)=(1+(e^-1)\varepsilon)^\mu The second derivative is K''(t)=(\varepsilon-(\varepsilon-1)e^t)^\mu\varepsilon e^t confirming that the first cumulant is and the second cumulant is . The constant random variables have . The binomial distributions have so that . The Poisson distributions have . The negative binomial distributions have so that . Note the analogy to the classification of
conic sections A conic section, conic or a quadratic curve is a curve obtained from a Conical surface, cone's surface intersecting a plane (mathematics), plane. The three types of conic section are the hyperbola, the parabola, and the ellipse; the circle is ...
by
eccentricity Eccentricity or eccentric may refer to: * Eccentricity (behavior), odd behavior on the part of a person, as opposed to being "normal" Mathematics, science and technology Mathematics * Off-Centre (geometry), center, in geometry * Eccentricity (g ...
: circles , ellipses , parabolas , hyperbolas .


Cumulants of some continuous probability distributions

* For the
normal distribution In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is f(x) = \frac ...
with
expected value In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
and
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
, the cumulant generating function is . The first and second derivatives of the cumulant generating function are and . The cumulants are , , and . The special case is a constant random variable . * The cumulants of the uniform distribution on the interval are , where is the th Bernoulli number. * The cumulants of the
exponential distribution In probability theory and statistics, the exponential distribution or negative exponential distribution is the probability distribution of the distance between events in a Poisson point process, i.e., a process in which events occur continuousl ...
with rate parameter are .


Some properties of the cumulant generating function

The cumulant generating function , if it exists, is infinitely differentiable and
convex Convex or convexity may refer to: Science and technology * Convex lens, in optics Mathematics * Convex set, containing the whole line segment that joins points ** Convex polygon, a polygon which encloses a convex set of points ** Convex polytop ...
, and passes through the origin. Its first derivative ranges monotonically in the open interval from the
infimum In mathematics, the infimum (abbreviated inf; : infima) of a subset S of a partially ordered set P is the greatest element in P that is less than or equal to each element of S, if such an element exists. If the infimum of S exists, it is unique ...
to the
supremum In mathematics, the infimum (abbreviated inf; : infima) of a subset S of a partially ordered set P is the greatest element in P that is less than or equal to each element of S, if such an element exists. If the infimum of S exists, it is unique, ...
of the support of the probability distribution, and its second derivative is strictly positive everywhere it is defined, except for the
degenerate distribution In probability theory, a degenerate distribution on a measure space (E, \mathcal, \mu) is a probability distribution whose support is a null set with respect to \mu. For instance, in the -dimensional space endowed with the Lebesgue measure, an ...
of a single point mass. The cumulant-generating function exists if and only if the tails of the distribution are majorized by an
exponential decay A quantity is subject to exponential decay if it decreases at a rate proportional to its current value. Symbolically, this process can be expressed by the following differential equation, where is the quantity and (lambda Lambda (; uppe ...
, that is, (''see
Big O notation Big ''O'' notation is a mathematical notation that describes the asymptotic analysis, limiting behavior of a function (mathematics), function when the Argument of a function, argument tends towards a particular value or infinity. Big O is a memb ...
'') \begin & \exists c>0,\,\, F(x)=O(e^), x\to-\infty; \text \\ pt& \exists d>0,\,\, 1-F(x)=O(e^),x\to+\infty; \end where F is the
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ever ...
. The cumulant-generating function will have vertical asymptote(s) at the negative
supremum In mathematics, the infimum (abbreviated inf; : infima) of a subset S of a partially ordered set P is the greatest element in P that is less than or equal to each element of S, if such an element exists. If the infimum of S exists, it is unique, ...
of such , if such a supremum exists, and at the
supremum In mathematics, the infimum (abbreviated inf; : infima) of a subset S of a partially ordered set P is the greatest element in P that is less than or equal to each element of S, if such an element exists. If the infimum of S exists, it is unique, ...
of such , if such a supremum exists, otherwise it will be defined for all real numbers. If the support of a random variable has finite upper or lower bounds, then its cumulant-generating function , if it exists, approaches
asymptote In analytic geometry, an asymptote () of a curve is a line such that the distance between the curve and the line approaches zero as one or both of the ''x'' or ''y'' coordinates tends to infinity. In projective geometry and related contexts, ...
(s) whose slope is equal to the supremum or infimum of the support, \begin y & =(t+1)\inf \operatorname X-\mu(X), \text \\ pty & =(t-1)\sup \operatornameX+\mu(X), \end respectively, lying above both these lines everywhere. (The
integral In mathematics, an integral is the continuous analog of a Summation, sum, which is used to calculate area, areas, volume, volumes, and their generalizations. Integration, the process of computing an integral, is one of the two fundamental oper ...
s \int_^0 \left \inf \operatornameX-K'(t)\right,dt, \qquad \int_^0 \left \inf \operatornameX-K'(t) \right,dt yield the -intercepts of these asymptotes, since .) For a shift of the distribution by , K_(t)=K_X(t)+ct. For a degenerate point mass at , the cumulant generating function is the straight line K_c(t)=ct, and more generally, K_=K_X+K_Y if and only if and are independent and their cumulant generating functions exist; ( subindependence and the existence of second moments sufficing to imply independence.) The natural exponential family of a distribution may be realized by shifting or translating , and adjusting it vertically so that it always passes through the origin: if is the pdf with cumulant generating function K(t)=\log M(t), and f, \theta is its natural exponential family, then f(x\mid\theta)=\frac1e^ f(x), and K(t\mid\theta)=K(t+\theta)-K(\theta). If is finite for a range then if then is analytic and infinitely differentiable for . Moreover for real and is strictly convex, and is strictly increasing.


Further properties of cumulants


A negative result

Given the results for the cumulants of the
normal distribution In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is f(x) = \frac ...
, it might be hoped to find families of distributions for which for some , with the lower-order cumulants (orders 3 to ) being non-zero. There are no such distributions. The underlying result here is that the cumulant generating function cannot be a finite-order polynomial of degree greater than 2.


Cumulants and moments

The moment generating function is given by: M(t) = 1+\sum_^\infty \frac = \exp \left(\sum_^\infty \frac\right) = \exp(K(t)). So the cumulant generating function is the logarithm of the moment generating function K(t) = \log M(t). The first cumulant is the
expected value In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
; the second and third cumulants are respectively the second and third central moments (the second central moment is the
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
); but the higher cumulants are neither moments nor central moments, but rather more complicated polynomial functions of the moments. The moments can be recovered in terms of cumulants by evaluating the th derivative of \exp(K(t)) at , \mu'_n = M^(0) = \left. \frac\_. Likewise, the cumulants can be recovered in terms of moments by evaluating the th derivative of \log M(t) at , \kappa_n = K^(0) = \left. \frac \_. The explicit expression for the th moment in terms of the first cumulants, and vice versa, can be obtained by using
Faà di Bruno's formula Faà di Bruno's formula is an identity in mathematics generalizing the chain rule to higher derivatives. It is named after , although he was not the first to state or prove the formula. In 1800, more than 50 years before Faà di Bruno, the French ...
for higher derivatives of composite functions. In general, we have \mu'_n = \sum_^n B_(\kappa_1,\ldots,\kappa_) \kappa_n = \sum_^n (-1)^ (k-1)! B_(\mu'_1, \ldots, \mu'_), where B_ are incomplete (or partial) Bell polynomials. In the like manner, if the mean is given by \mu, the central moment generating function is given by C(t) = \operatorname ^= e^ M(t) = \exp(K(t) - \mu t), and the th central moment is obtained in terms of cumulants as \mu_n = C^(0) = \left. \frac \exp (K(t) - \mu t) \_ = \sum_^n B_(0,\kappa_2,\ldots,\kappa_). Also, for , the th cumulant in terms of the central moments is \begin \kappa_n & = K^(0) = \left. \frac (\log C(t) + \mu t) \_ \\ pt& = \sum_^n (-1)^ (k-1)! B_(0,\mu_2,\ldots,\mu_). \end The th moment is an th-degree polynomial in the first cumulants. The first few expressions are: \begin \mu'_1 = & \kappa_1 \\ pt\mu'_2 = & \kappa_2+\kappa_1^2 \\ pt\mu'_3 = & \kappa_3+3\kappa_2\kappa_1+\kappa_1^3 \\ pt\mu'_4 = & \kappa_4 + 4\kappa_3\kappa_1 + 3\kappa_2^2 + 6\kappa_2\kappa_1^2 + \kappa_1^4 \\ pt\mu'_5 = & \kappa_5+5\kappa_4\kappa_1+10\kappa_3\kappa_2 + 10\kappa_3\kappa_1^2 + 15\kappa_2^2\kappa_1 + 10\kappa_2\kappa_1^3 + \kappa_1^5 \\ pt\mu'_6 = & \kappa_6 + 6\kappa_5\kappa_1 + 15\kappa_4\kappa_2 + 15\kappa_4\kappa_1^2 + 10\kappa_3^2 + 60\kappa_3\kappa_2\kappa_1 + 20\kappa_3\kappa_1^3 \\ & + 15\kappa_2^3 + 45\kappa_2^2\kappa_1^2 + 15\kappa_2\kappa_1^4 + \kappa_1^6. \end The "prime" distinguishes the moments from the central moments . To express the ''central'' moments as functions of the cumulants, just drop from these polynomials all terms in which appears as a factor: \begin \mu_1 & =0 \\ pt\mu_2 & =\kappa_2 \\ pt\mu_3 & =\kappa_3 \\ pt\mu_4 & =\kappa_4+3\kappa_2^2 \\ pt\mu_5 & =\kappa_5+10\kappa_3\kappa_2 \\ pt\mu_6 & =\kappa_6+15\kappa_4\kappa_2+10\kappa_3^2+15\kappa_2^3. \end Similarly, the th cumulant is an th-degree polynomial in the first non-central moments. The first few expressions are: \begin \kappa_1 = & \mu'_1 \\ pt\kappa_2 = & \mu'_2-^2 \\ pt\kappa_3 = & \mu'_3-3\mu'_2\mu'_1+2^3 \\ pt\kappa_4 = & \mu'_4-4\mu'_3\mu'_1-3^2+12\mu'_2^2-6^4 \\ pt\kappa_5 = & \mu'_5-5\mu'_4\mu'_1-10\mu'_3\mu'_2 + 20\mu'_3^2 + 30^2\mu'_1-60\mu'_2^3 + 24^5 \\ pt\kappa_6 = & \mu'_6-6\mu'_5\mu'_1-15\mu'_4\mu'_2+30\mu'_4^2-10^2 + 120\mu'_3\mu'_2\mu'_1 \\ & - 120\mu'_3^3 + 30^3 - 270^2 ^2+360\mu'_2^4-120^6\,. \end In general, the cumulant is the determinant of a matrix: \kappa_l = (-1)^ \left, \begin \mu'_1 & 1 & 0 & 0 & 0 & 0 & \ldots & 0 \\ \mu'_2 & \mu'_1 & 1 & 0 & 0 & 0 & \ldots & 0 \\ \mu'_3 & \mu'_2 & \left(\begin 2 \\ 1 \end\right) \mu'_1 & 1 & 0 & 0 & \ldots & 0 \\ \mu'_4 & \mu'_3 & \left(\begin 3 \\ 1 \end\right) \mu'_2 & \left(\begin 3 \\ 2 \end\right) \mu'_1 & 1 & 0 & \ldots & 0 \\ \mu'_5 & \mu'_4 & \left(\begin 4 \\ 1 \end\right) \mu'_3 & \left(\begin 4 \\ 2 \end\right) \mu'_2 & \left(\begin 4 \\ 3 \end\right) \mu'_1 & 1 & \ldots & 0 \\ \vdots & \vdots & \vdots & \vdots & \vdots & \ddots & \ddots & \vdots \\ \mu'_ & \mu'_ & \ldots & \ldots & \ldots & \ldots & \ddots & 1 \\ \mu'_l & \mu'_ & \ldots & \ldots & \ldots & \ldots & \ldots & \left(\begin l-1 \\ l-2 \end\right) \mu'_1 \end\ To express the cumulants for as functions of the central moments, drop from these polynomials all terms in which μ'1 appears as a factor: \kappa_2=\mu_2\, \kappa_3=\mu_3\, \kappa_4=\mu_4-3^2\, \kappa_5=\mu_5-10\mu_3\mu_2\, \kappa_6=\mu_6-15\mu_4\mu_2-10^2+30^3\,. The cumulants can be related to the moments by differentiating the relationship with respect to , giving , which conveniently contains no exponentials or logarithms. Equating the coefficient of on the left and right sides and using gives the following formulas for : \begin \mu'_1 = & \kappa_1 \\ pt\mu'_2 = & \kappa_1\mu'_1+\kappa_2 \\ pt\mu'_3 = & \kappa_1\mu'_2+2\kappa_2\mu'_1+\kappa_3 \\ pt\mu'_4 = & \kappa_1\mu'_3+3\kappa_2\mu'_2+3\kappa_3\mu'_1+\kappa_4 \\ pt\mu'_5 = & \kappa_1\mu'_4+4\kappa_2\mu'_3+6\kappa_3\mu'_2+4\kappa_4\mu'_1+\kappa_5 \\ pt\mu'_6 = & \kappa_1\mu'_5+5\kappa_2\mu'_4+10\kappa_3\mu'_3+10\kappa_4\mu'_2+5\kappa_5\mu'_1+\kappa_6 \\ pt\mu'_n = & \sum_^\kappa_m \mu'_ + \kappa_n\,. \end These allow either \kappa_n or \mu'_n to be computed from the other using knowledge of the lower-order cumulants and moments. The corresponding formulas for the central moments \mu_n for n \ge 2 are formed from these formulas by setting \mu'_1 = \kappa_1 = 0 and replacing each \mu'_n with \mu_n for n \ge 2: \begin \mu_2 = & \kappa_2 \\ pt\mu_3 = & \kappa_3 \\ pt\mu_n = & \sum_^\kappa_m \mu_ + \kappa_n\,. \end


Cumulants and set-partitions

These polynomials have a remarkable
combinatorial Combinatorics is an area of mathematics primarily concerned with counting, both as a means and as an end to obtaining results, and certain properties of finite structures. It is closely related to many other areas of mathematics and has many ...
interpretation: the coefficients count certain partitions of sets. A general form of these polynomials is \mu'_n=\sum_ \prod_ \kappa_ where * runs through the list of all partitions of a set of size ; * "" means is one of the "blocks" into which the set is partitioned; and * is the size of the set . Thus each
monomial In mathematics, a monomial is, roughly speaking, a polynomial which has only one term. Two definitions of a monomial may be encountered: # A monomial, also called a power product or primitive monomial, is a product of powers of variables with n ...
is a constant times a product of cumulants in which the sum of the indices is (e.g., in the term , the sum of the indices is 3 + 2 + 2 + 1 = 8; this appears in the polynomial that expresses the 8th moment as a function of the first eight cumulants). A partition of the
integer An integer is the number zero (0), a positive natural number (1, 2, 3, ...), or the negation of a positive natural number (−1, −2, −3, ...). The negations or additive inverses of the positive natural numbers are referred to as negative in ...
corresponds to each term. The ''coefficient'' in each term is the number of partitions of a set of members that collapse to that partition of the integer when the members of the set become indistinguishable.


Cumulants and combinatorics

Further connection between cumulants and combinatorics can be found in the work of
Gian-Carlo Rota Gian-Carlo Rota (April 27, 1932 – April 18, 1999) was an Italian-American mathematician and philosopher. He spent most of his career at the Massachusetts Institute of Technology, where he worked in combinatorics, functional analysis, proba ...
, where links to
invariant theory Invariant theory is a branch of abstract algebra dealing with actions of groups on algebraic varieties, such as vector spaces, from the point of view of their effect on functions. Classically, the theory dealt with the question of explicit descr ...
,
symmetric function In mathematics, a function of n variables is symmetric if its value is the same no matter the order of its arguments. For example, a function f\left(x_1,x_2\right) of two arguments is a symmetric function if and only if f\left(x_1,x_2\right) = f\ ...
s, and binomial sequences are studied via umbral calculus.


Joint cumulants

The joint cumulant of several random variables is defined as the coefficient in the Maclaurin series of the multivariate cumulant generating function, see Section 3.1 in, G(t_1,\dots,t_n)=\log \mathrm(\mathrm^) =\sum_ \kappa_ \frac \,. Note that \kappa_ = \left.\left(\frac\right)^ \cdots \left(\frac\right)^ G(t_1,\dots,t_n) \_\,, and, in particular \kappa(X_1,\ldots,X_n) = \left. \frac G(t_1,\dots,t_n) \_\,. As with a single variable, the generating function and cumulant can instead be defined via H(t_1,\dots,t_n) =\log \mathrm(\mathrm^) =\sum_ \kappa_ i^ \frac\,, in which case \kappa_ = (-i)^ \left.\left(\frac\right)^ \cdots \left(\frac\right)^ H(t_1,\dots,t_n) \_\,, and \kappa(X_1,\ldots,X_n) = \left. (-i)^ \frac H(t_1,\dots,t_n) \_\,.


Repeated random variables and relation between the coefficients ''κ''''k''1, ..., ''k''''n''

Observe that \kappa_ (X_1,\ldots,X_n) can also be written as \kappa_ = \left. \frac \cdots \frac G\left(\sum_^t_,\dots,\sum_^t_\right) \_, from which we conclude that \kappa_ (X_1,\ldots,X_n) = \kappa_ ( \underbrace_, \ldots , \underbrace_ ) . For example \kappa_(X,Y,Z) = \kappa(X,X,Z),\, and \kappa_(X,Y,Z,T) = \kappa_(Z) = \kappa(\underbrace_) .\, In particular, the last equality shows that the cumulants of a single random variable are the joint cumulants of multiple copies of that random variable.


Relation with mixed moments

The joint cumulant of random variables can be expressed as an alternate sum of products of their mixed moments, see Equation (3.2.7) in, \kappa(X_1,\dots,X_n)=\sum_\pi (, \pi, -1)!(-1)^\prod_E\left(\prod_X_i\right) where  runs through the list of all partitions of ; where  runs through the list of all blocks of the partition ; and where  is the number of parts in the partition. For example, \kappa(X)=\operatorname E(X), is the expected value of X, \kappa(X,Y)=\operatorname E(XY) - \operatorname E(X) \operatorname E(Y), is the
covariance In probability theory and statistics, covariance is a measure of the joint variability of two random variables. The sign of the covariance, therefore, shows the tendency in the linear relationship between the variables. If greater values of one ...
of X and Y, and \kappa(X,Y,Z)=\operatorname E(XYZ) - \operatorname E(XY) \operatorname E(Z) - \operatorname E(XZ) \operatorname E(Y) - \operatorname E(YZ) \operatorname E(X) + 2\operatorname E(X)\operatorname E(Y)\operatorname E(Z).\, For zero-mean random variables X_1,\ldots,X_n, any mixed moment of the form \prod_ E\left(\prod_ X_i\right) vanishes if \pi is a partition of \ which contains a singleton B=\. Hence, the expression of their joint cumulant in terms of mixed moments simplifies. For example, if X,Y,Z,W are zero mean random variables, we have \kappa(X,Y,Z) = \operatorname E(XYZ).\, \kappa(X,Y,Z,W) = \operatorname E(XYZW) - \operatorname E(XY) \operatorname E(ZW) - \operatorname E(XZ) \operatorname E(YW) - \operatorname E(XW) \operatorname E(YZ).\, More generally, any coefficient of the Maclaurin series can also be expressed in terms of mixed moments, although there are no concise formulae. Indeed, as noted above, one can write it as a joint cumulant by repeating random variables appropriately, and then apply the above formula to express it in terms of mixed moments. For example \kappa_(X,Y,Z) = \kappa(X,X,Z)=\operatorname E(X^2Z) -2\operatorname E(XZ)\operatorname E(X) - \operatorname E(X^2)\operatorname E(Z) + 2\operatorname E(X)^2\operatorname E(Z).\, If some of the random variables are independent of all of the others, then any cumulant involving two (or more) independent random variables is zero. The combinatorial meaning of the expression of mixed moments in terms of cumulants is easier to understand than that of cumulants in terms of mixed moments, see Equation (3.2.6) in: \operatorname E(X_1\cdots X_n)=\sum_\pi\prod_\kappa(X_i : i \in B). For example: \operatorname E(XYZ) = \kappa(X,Y,Z) + \kappa(X,Y)\kappa(Z) + \kappa(X,Z)\kappa(Y) + \kappa(Y,Z)\kappa(X) + \kappa(X)\kappa(Y)\kappa(Z).\,


Further properties

Another important property of joint cumulants is multilinearity: \kappa(X+Y,Z_1,Z_2,\dots) = \kappa(X,Z_1,Z_2,\ldots) + \kappa(Y,Z_1,Z_2,\ldots).\, Just as the second cumulant is the variance, the joint cumulant of just two random variables is the
covariance In probability theory and statistics, covariance is a measure of the joint variability of two random variables. The sign of the covariance, therefore, shows the tendency in the linear relationship between the variables. If greater values of one ...
. The familiar identity \operatorname(X+Y) = \operatorname(X) + 2\operatorname(X,Y) + \operatorname(Y)\, generalizes to cumulants: \kappa_n(X+Y)=\sum_^n \kappa( \, \underbrace_j, \underbrace_\,).\,


Conditional cumulants and the law of total cumulance

The
law of total expectation The proposition in probability theory known as the law of total expectation, the law of iterated expectations (LIE), Adam's law, the tower rule, and the smoothing property of conditional expectation, among other names, states that if X is a random ...
and the law of total variance generalize naturally to conditional cumulants. The case , expressed in the language of (central) moments rather than that of cumulants, says \mu_3(X) = \operatorname E(\mu_3(X\mid Y)) + \mu_3(\operatorname E(X\mid Y)) + 3 \operatorname(\operatorname E(X\mid Y), \operatorname (X\mid Y)). In general, \kappa(X_1,\dots,X_n)=\sum_\pi \kappa(\kappa(X_\mid Y), \dots, \kappa(X_\mid Y)) where * the sum is over all partitions  of the set of indices, and * 1, ..., b are all of the "blocks" of the partition ; the expression indicates that the joint cumulant of the random variables whose indices are in that block of the partition.


Conditional cumulants and the conditional expectation

For certain settings, a derivative identity can be established between the conditional cumulant and the conditional expectation. For example, suppose that where is standard normal independent of , then for any it holds that \kappa_(X\mid Y=y) = \frac\operatorname E(X\mid Y = y), \, n \in \mathbb, \, y \in \mathbb. The results can also be extended to the exponential family.


Relation to statistical physics

In
statistical physics In physics, statistical mechanics is a mathematical framework that applies statistical methods and probability theory to large assemblies of microscopic entities. Sometimes called statistical physics or statistical thermodynamics, its applicati ...
many extensive quantities – that is quantities that are proportional to the volume or size of a given system – are related to cumulants of random variables. The deep connection is that in a large system an extensive quantity like the energy or number of particles can be thought of as the sum of (say) the energy associated with a number of nearly independent regions. The fact that the cumulants of these nearly independent random variables will (nearly) add make it reasonable that extensive quantities should be expected to be related to cumulants. A system in equilibrium with a thermal bath at temperature have a fluctuating internal energy , which can be considered a random variable drawn from a distribution E\sim p(E). The partition function of the system is Z(\beta) = \sum_i e^ , where =  and is the
Boltzmann constant The Boltzmann constant ( or ) is the proportionality factor that relates the average relative thermal energy of particles in a ideal gas, gas with the thermodynamic temperature of the gas. It occurs in the definitions of the kelvin (K) and the ...
and the notation \langle A \rangle has been used rather than \operatorname /math> for the expectation value to avoid confusion with the energy, . Hence the first and second cumulant for the energy give the average energy and heat capacity. \begin \langle E \rangle_c & = \frac = \langle E \rangle \\ pt\langle E^2 \rangle_c & = \frac = k T^2 \frac = kT^2C \end The
Helmholtz free energy In thermodynamics, the Helmholtz free energy (or Helmholtz energy) is a thermodynamic potential that measures the useful work obtainable from a closed thermodynamic system at a constant temperature ( isothermal). The change in the Helmholtz ene ...
expressed in terms of F(\beta) = -\beta^\log Z(\beta) \, further connects thermodynamic quantities with cumulant generating function for the energy. Thermodynamics properties that are derivatives of the free energy, such as its
internal energy The internal energy of a thermodynamic system is the energy of the system as a state function, measured as the quantity of energy necessary to bring the system from its standard internal state to its present internal state of interest, accoun ...
,
entropy Entropy is a scientific concept, most commonly associated with states of disorder, randomness, or uncertainty. The term and the concept are used in diverse fields, from classical thermodynamics, where it was first recognized, to the micros ...
, and
specific heat In thermodynamics, the specific heat capacity (symbol ) of a substance is the amount of heat that must be added to one unit of mass of the substance in order to cause an increase of one unit in temperature. It is also referred to as massic heat ...
capacity, all can be readily expressed in terms of these cumulants. Other free energy can be a function of other variables such as the magnetic field or chemical potential \mu, e.g. \Omega=-\beta^\log(\langle \exp(-\beta E -\beta\mu N) \rangle),\, where is the number of particles and \Omega is the grand potential. Again the close relationship between the definition of the free energy and the cumulant generating function implies that various derivatives of this free energy can be written in terms of joint cumulants of and .


History

The history of cumulants is discussed by Anders Hald. Cumulants were first introduced by Thorvald N. Thiele, in 1889, who called them ''semi-invariants''. They were first called ''cumulants'' in a 1932 paper by
Ronald Fisher Sir Ronald Aylmer Fisher (17 February 1890 – 29 July 1962) was a British polymath who was active as a mathematician, statistician, biologist, geneticist, and academic. For his work in statistics, he has been described as "a genius who a ...
and John Wishart. Fisher was publicly reminded of Thiele's work by Neyman, who also notes previous published citations of Thiele brought to Fisher's attention. Stephen Stigler has said that the name ''cumulant'' was suggested to Fisher in a letter from
Harold Hotelling Harold Hotelling (; September 29, 1895 – December 26, 1973) was an American mathematical statistician and an influential economic theorist, known for Hotelling's law, Hotelling's lemma, and Hotelling's rule in economics, as well as Hotelling ...
. In a paper published in 1929, Fisher had called them ''cumulative moment functions''. The partition function in statistical physics was introduced by Josiah Willard Gibbs in 1901. The free energy is often called Gibbs free energy. In
statistical mechanics In physics, statistical mechanics is a mathematical framework that applies statistical methods and probability theory to large assemblies of microscopic entities. Sometimes called statistical physics or statistical thermodynamics, its applicati ...
, cumulants are also known as Ursell functions relating to a publication in 1927.


Cumulants in generalized settings


Formal cumulants

More generally, the cumulants of a sequence , not necessarily the moments of any probability distribution, are, by definition, 1+\sum_^\infty \frac = \exp \left( \sum_^\infty \frac \right) , where the values of for are found formally, i.e., by algebra alone, in disregard of questions of whether any series converges. All of the difficulties of the "problem of cumulants" are absent when one works formally. The simplest example is that the second cumulant of a probability distribution must always be nonnegative, and is zero only if all of the higher cumulants are zero. Formal cumulants are subject to no such constraints.


Bell numbers

In
combinatorics Combinatorics is an area of mathematics primarily concerned with counting, both as a means and as an end to obtaining results, and certain properties of finite structures. It is closely related to many other areas of mathematics and has many ...
, the th Bell number is the number of partitions of a set of size . All of the cumulants of the sequence of Bell numbers are equal to 1. The Bell numbers are the moments of the Poisson distribution with expected value 1.


Cumulants of a polynomial sequence of binomial type

For any sequence of
scalars Scalar may refer to: *Scalar (mathematics), an element of a field, which is used to define a vector space, usually the field of real numbers *Scalar (physics), a physical quantity that can be described by a single element of a number field such a ...
in a field of characteristic zero, being considered formal cumulants, there is a corresponding sequence of formal moments, given by the polynomials above. For those polynomials, construct a
polynomial sequence In mathematics, a polynomial sequence is a sequence of polynomials indexed by the nonnegative integers 0, 1, 2, 3, ..., in which each index is equal to the degree of the corresponding polynomial. Polynomial sequences are a topic of interest in ...
in the following way. Out of the polynomial \begin \mu'_6 = & \kappa_6 + 6\kappa_5\kappa_1 + 15\kappa_4\kappa_2 + 15\kappa_4\kappa_1^2 + 10\kappa_3^2+60\kappa_3\kappa_2\kappa_1 + 20\kappa_3\kappa_1^3 \\ & + 15\kappa_2^3 + 45\kappa_2^2\kappa_1^2 + 15\kappa_2\kappa_1^4 + \kappa_1^6 \end make a new polynomial in these plus one additional variable : \begin p_6(x) = & \kappa_6 \,x + (6\kappa_5\kappa_1 + 15\kappa_4\kappa_2 + 10\kappa_3^2)\,x^2 + (15\kappa_4\kappa_1^2 + 60\kappa_3\kappa_2\kappa_1 + 15\kappa_2^3)\,x^3 \\ & + (45\kappa_2^2\kappa_1^2)\,x^4+(15\kappa_2\kappa_1^4)\,x^5 +(\kappa_1^6)\,x^6, \end and then generalize the pattern. The pattern is that the numbers of blocks in the aforementioned partitions are the exponents on . Each coefficient is a polynomial in the cumulants; these are the Bell polynomials, named after Eric Temple Bell. This sequence of polynomials is of binomial type. In fact, no other sequences of binomial type exist; every polynomial sequence of binomial type is completely determined by its sequence of formal cumulants.


Free cumulants

In the above moment-cumulant formula\ \operatorname E(X_1\cdots X_n)=\sum_\pi\prod_\kappa(X_i : i\in B) for joint cumulants, one sums over ''all'' partitions of the set . If instead, one sums only over the noncrossing partitions, then, by solving these formulae for the \kappa in terms of the moments, one gets free cumulants rather than conventional cumulants treated above. These free cumulants were introduced by Roland Speicher and play a central role in free probability theory. In that theory, rather than considering
independence Independence is a condition of a nation, country, or state, in which residents and population, or some portion thereof, exercise self-government, and usually sovereignty, over its territory. The opposite of independence is the status of ...
of
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
s, defined in terms of tensor products of algebras of random variables, one considers instead free independence of random variables, defined in terms of
free product In mathematics, specifically group theory, the free product is an operation that takes two groups ''G'' and ''H'' and constructs a new The result contains both ''G'' and ''H'' as subgroups, is generated by the elements of these subgroups, an ...
s of algebras. The ordinary cumulants of degree higher than 2 of the
normal distribution In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is f(x) = \frac ...
are zero. The ''free'' cumulants of degree higher than 2 of the Wigner semicircle distribution are zero. This is one respect in which the role of the Wigner distribution in free probability theory is analogous to that of the normal distribution in conventional probability theory.


See also

* Entropic value at risk * Cumulant generating function from a multiset * Cornish–Fisher expansion * Edgeworth expansion * Polykay * k-statistic, a minimum-variance
unbiased estimator In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In stat ...
of a cumulant * Ursell function * Total position spread tensor as an application of cumulants to analyse the electronic
wave function In quantum physics, a wave function (or wavefunction) is a mathematical description of the quantum state of an isolated quantum system. The most common symbols for a wave function are the Greek letters and (lower-case and capital psi (letter) ...
in
quantum chemistry Quantum chemistry, also called molecular quantum mechanics, is a branch of physical chemistry focused on the application of quantum mechanics to chemical systems, particularly towards the quantum-mechanical calculation of electronic contributions ...
.


References


External links

*
cumulant
on th

{{Theory of probability distributions Moments (mathematics)