In
probability theory
Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
and
statistics, the Hermite distribution, named after
Charles Hermite
Charles Hermite () FRS FRSE MIAS (24 December 1822 – 14 January 1901) was a French mathematician who did research concerning number theory, quadratic forms, invariant theory, orthogonal polynomials, elliptic functions, and algebra.
Herm ...
, is a
discrete probability distribution
In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon i ...
used to model ''count data'' with more than one parameter. This distribution is flexible in terms of its ability to allow a moderate
over-dispersion
In statistics, overdispersion is the presence of greater variability (statistical dispersion) in a data set than would be expected based on a given statistical model.
A common task in applied statistics is choosing a parametric model to fit a g ...
in the data.
The authors Kemp and Kemp
have called it "Hermite distribution" from the fact its
probability function Probability function may refer to:
* Probability distribution
* Probability axioms, which define a probability function
* Probability measure
In mathematics, a probability measure is a real-valued function defined on a set of events in a prob ...
and the
moment generating function
In probability theory and statistics, the moment-generating function of a real-valued random variable is an alternative specification of its probability distribution. Thus, it provides the basis of an alternative route to analytical results compar ...
can be expressed in terms of the coefficients of (modified)
Hermite polynomials
In mathematics, the Hermite polynomials are a classical orthogonal polynomial sequence.
The polynomials arise in:
* signal processing as Hermitian wavelets for wavelet transform analysis
* probability, such as the Edgeworth series, as well ...
.
History
The distribution first appeared in the paper ''Applications of Mathematics to Medical Problems'',
by
Anderson Gray McKendrick
Lt Col Anderson Gray McKendrick DSc FRSE (8 September 1876 – 30 May 1943) was a Scottish military physician and epidemiologist who pioneered the use of mathematical methods in epidemiology. Irwin (see below) commented on the quality of his wor ...
in 1926. In this work the author explains several mathematical methods that can be applied to medical research. In one of this methods he considered the
bivariate Poisson distribution and showed that the distribution of the sum of two correlated Poisson variables follow a distribution that later would be known as Hermite distribution.
As a practical application, McKendrick considered the distribution of counts of
bacteria
Bacteria (; singular: bacterium) are ubiquitous, mostly free-living organisms often consisting of one biological cell. They constitute a large domain of prokaryotic microorganisms. Typically a few micrometres in length, bacteria were am ...
in
leucocytes
White blood cells, also called leukocytes or leucocytes, are the cells of the immune system that are involved in protecting the body against both infectious disease and foreign invaders. All white blood cells are produced and derived from multi ...
. Using the
method of moments he fitted the data with the Hermite distribution and found the model more satisfactory than fitting it with a
Poisson distribution
In probability theory and statistics, the Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time or space if these events occur with a known ...
.
The distribution was formally introduced and published by C. D. Kemp and Adrienne W. Kemp in 1965 in their work ''Some Properties of ‘Hermite’ Distribution''. The work is focused on the properties of this distribution for instance a necessary condition on the parameters and their
maximum likelihood estimators (MLE), the analysis of the
probability generating function In probability theory, the probability generating function of a discrete random variable is a power series representation (the generating function) of the probability mass function of the random variable. Probability generating functions are often ...
(PGF) and how it can be expressed in terms of the coefficients of (modified)
Hermite polynomials
In mathematics, the Hermite polynomials are a classical orthogonal polynomial sequence.
The polynomials arise in:
* signal processing as Hermitian wavelets for wavelet transform analysis
* probability, such as the Edgeworth series, as well ...
. An example they have used in this publication is the distribution of counts of bacteria in leucocytes that used McKendrick but Kemp and Kemp estimate the model using the
maximum likelihood
In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed sta ...
method.
Hermite distribution is a special case of discrete
compound Poisson distribution
In probability theory, a compound Poisson distribution is the probability distribution of the sum of a number of independent identically-distributed random variables, where the number of terms to be added is itself a Poisson-distributed variable. T ...
with only two parameters.
[Johnson, N.L., Kemp, A.W., and Kotz, S. (2005) Univariate Discrete Distributions, 3rd Edition, Wiley, .]
The same authors published in 1966 the paper ''An alternative Derivation of the Hermite Distribution''.
In this work established that the Hermite distribution can be obtained formally by combining a
Poisson distribution
In probability theory and statistics, the Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time or space if these events occur with a known ...
with a
normal distribution
In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is
:
f(x) = \frac e^
The parameter \mu i ...
.
In 1971, Y. C. Patel
did a comparative study of various estimation procedures for the Hermite distribution in his doctoral thesis. It included maximum likelihood, moment estimators, mean and zero frequency estimators and the method of even points.
In 1974, Gupta and Jain
did a research on a generalized form of Hermite distribution.
Definition
Probability mass function
Let ''X''
1 and ''X''
2 be two independent Poisson variables with parameters ''a''
1 and ''a''
2. The
probability distribution
In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomeno ...
of the
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the p ...
''Y'' = ''X''
1 + 2''X''
2 is the Hermite distribution with parameters ''a''
1 and ''a''
2 and
probability mass function
In probability and statistics, a probability mass function is a function that gives the probability that a discrete random variable is exactly equal to some value. Sometimes it is also known as the discrete density function. The probability mass ...
is given by
:
where
* ''n'' = 0, 1, 2, ...
* ''a''
1, ''a''
2 ≥ 0.
* (''n'' − 2''j'')! and ''j''! are the
factorial
In mathematics, the factorial of a non-negative denoted is the product of all positive integers less than or equal The factorial also equals the product of n with the next smaller factorial:
\begin
n! &= n \times (n-1) \times (n-2) ...
s of (''n'' − 2''j'') and ''j'', respectively.
*
is the integer part of ''n''/2.
The
probability generating function In probability theory, the probability generating function of a discrete random variable is a power series representation (the generating function) of the probability mass function of the random variable. Probability generating functions are often ...
of the probability mass is,
[
:
]
Notation
When a random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the p ...
''Y'' = ''X''1 + 2''X''2 is distributed by an Hermite distribution, where ''X''1 and ''X''2 are two independent Poisson variables with parameters ''a''1 and ''a''2, we write
:
Properties
Moment and cumulant generating functions
The moment generating function
In probability theory and statistics, the moment-generating function of a real-valued random variable is an alternative specification of its probability distribution. Thus, it provides the basis of an alternative route to analytical results compar ...
of a random variable ''X'' is defined as the expected value of ''e''''t'', as a function of the real parameter ''t''. For an Hermite distribution with parameters ''X''1 and ''X''2, the moment generating function exists and is equal to
:
The cumulant generating function
In probability theory and statistics, the cumulants of a probability distribution are a set of quantities that provide an alternative to the '' moments'' of the distribution. Any two probability distributions whose moments are identical will hav ...
is the logarithm of the moment generating function and is equal to [
:
If we consider the coefficient of (''it'')''r''''r''! in the expansion of ''K''(''t'') we obtain the ''r''-cumulant
:
Hence the ]mean
There are several kinds of mean in mathematics, especially in statistics. Each mean serves to summarize a given group of data, often to better understand the overall value ( magnitude and sign) of a given data set.
For a data set, the '' ari ...
and the succeeding three moments about it are
Skewness
The skewness
In probability theory and statistics, skewness is a measure of the asymmetry of the probability distribution of a real-valued random variable about its mean. The skewness value can be positive, zero, negative, or undefined.
For a unimo ...
is the third moment centered around the mean divided by the 3/2 power of the standard deviation, and for the hermite distribution is,[
:
*Always , so the mass of the distribution is concentrated on the left.
]
Kurtosis
The kurtosis
In probability theory and statistics, kurtosis (from el, κυρτός, ''kyrtos'' or ''kurtos'', meaning "curved, arching") is a measure of the "tailedness" of the probability distribution of a real-valued random variable. Like skewness, kur ...
is the fourth moment centered around the mean, divided by the square of the variance
In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of number ...
, and for the Hermite distribution is,[
:
The ]excess kurtosis
In probability theory and statistics, kurtosis (from el, κυρτός, ''kyrtos'' or ''kurtos'', meaning "curved, arching") is a measure of the "tailedness" of the probability distribution of a real-valued random variable. Like skewness, kurtos ...
is just a correction to make the kurtosis of the normal distribution equal to zero, and it is the following,
:
*Always , or the distribution has a high acute peak around the mean and fatter tails.
Characteristic function
In a discrete distribution the characteristic function In mathematics, the term "characteristic function" can refer to any of several distinct concepts:
* The indicator function of a subset, that is the function
::\mathbf_A\colon X \to \,
:which for a given subset ''A'' of ''X'', has value 1 at point ...
of any real-valued random variable is defined as the expected value
In probability theory, the expected value (also called expectation, expectancy, mathematical expectation, mean, average, or first moment) is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a ...
of , where ''i'' is the imaginary unit and ''t'' ∈ ''R''
: