In the mathematical theory of
random processes
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables. Stochastic processes are widely used as mathematical models of systems and phenomena that appea ...
, the Markov chain central limit theorem has a conclusion somewhat similar in form to that of the classic
central limit theorem
In probability theory, the central limit theorem (CLT) establishes that, in many situations, when independent random variables are summed up, their properly normalized sum tends toward a normal distribution even if the original variables thems ...
(CLT) of probability theory, but the quantity in the role taken by the variance in the classic CLT has a more complicated definition. See also the general form of
Bienaymé's identity.
Statement
Suppose that:
* the sequence
of
random element In probability theory, random element is a generalization of the concept of random variable to more complicated spaces than the simple real line. The concept was introduced by who commented that the “development of probability theory and expansio ...
s of some set is a
Markov chain
A Markov chain or Markov process is a stochastic model describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event. Informally, this may be thought of as, "What happen ...
that has a
stationary probability distribution Stationary distribution may refer to:
* A special distribution for a Markov chain such that if the chain starts with its stationary distribution, the marginal distribution of all states at any time will always be the stationary distribution. Assum ...
; and
* the initial distribution of the process, i.e. the distribution of
, is the stationary distribution, so that
are identically distributed. In the classic central limit theorem these random variables would be assumed to be
independent
Independent or Independents may refer to:
Arts, entertainment, and media Artist groups
* Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s
* Independe ...
, but here we have only the weaker assumption that the process has the
Markov property
In probability theory and statistics, the term Markov property refers to the memoryless property of a stochastic process. It is named after the Russian mathematician Andrey Markov. The term strong Markov property is similar to the Markov prop ...
; and
*
is some (measurable) real-valued function for which
Now let
:
Then as
we have
[Geyer, Charles J. (2011). Introduction to Markov Chain Monte Carlo. In ''Handbook of MarkovChain Monte Carlo''. Edited by S. P. Brooks, A. E. Gelman, G. L. Jones, and X. L. Meng. Chapman & Hall/CRC, Boca Raton, FL, Section 1.8. http://www.mcmchandbook.net/HandbookChapter1.pdf]
:
where the decorated arrow indicates
convergence in distribution
In probability theory, there exist several different notions of convergence of random variables. The convergence of sequences of random variables to some limit random variable is an important concept in probability theory, and its applications to ...
.
This means that
:
where ~ means "is distributed like".
Monte Carlo Setting
The Markov chain central limit theorem can be guaranteed for functionals of general state space Markov chains under certain conditions. In particular, this can be done with a focus on Monte Carlo settings. An example of the application in a MCMC (Markov Chain Monte Carlo) setting is the following:
Consider a simple
hard spheres
Hard spheres are widely used as model particles in the statistical mechanical theory of fluids and solids. They are defined simply as impenetrable spheres that cannot overlap in space. They mimic the extremely strong ("infinitely elastic bouncing" ...
model on a grid. Suppose
. A proper configuration on
consists of coloring each point either black or white in such a way that no two adjacent points are white. Let
denote the set of all proper configurations on
,
be the total number of proper configurations and π be the uniform distribution on
so that each proper configuration is equally likely. Suppose our goal is to calculate the typical number of white points in a proper configuration; that is, if
is the number of white points in
then we want the value of
If
and
are even moderately large then we will have to resort to an approximation to
. Consider the following Markov chain on
. Fix
and set
where
is an arbitrary proper configuration. Randomly choose a point
and independently draw
. If
and all of the adjacent points are black then color
white leaving all other points alone. Otherwise, color
black and leave all other points alone. Call the resulting configuration
. Continuing in this fashion yields a Harris ergodic Markov chain
having
as its invariant distribution. It is now a simple matter to estimate
with
. Also, since
is finite (albeit potentially large) it is well known that
will converge exponentially fast to
which implies that a CLT holds for
.
Implications
Not taking into account the additional terms in the variance which stem from correlations (e.g. serial correlations in markov chain monte carlo simulations) can result in the problem of
pseudoreplication
Pseudoreplication (sometimes unit of analysis error) has many definitions. Pseudoreplication was originally defined in 1984 by Stuart H. Hurlbert as the use of inferential statistics to test for treatment effects with data from experiments where ...
when computing e.g. the
confidence interval
In frequentist statistics, a confidence interval (CI) is a range of estimates for an unknown parameter. A confidence interval is computed at a designated ''confidence level''; the 95% confidence level is most common, but other levels, such as ...
s for the
sample mean
The sample mean (or "empirical mean") and the sample covariance are statistics computed from a sample of data on one or more random variables.
The sample mean is the average value (or mean value) of a sample of numbers taken from a larger po ...
.
References
{{reflist
Sources
* Gordin, M. I. and Lifšic, B. A. (1978). "Central limit theorem for stationary Markov processes." ''Soviet Mathematics, Doklady'', 19, 392–394. (English translation of Russian original).
* Geyer, Charles J. (2011). "Introduction to MCMC." In ''Handbook of Markov Chain Monte Carlo'', edited by S. P. Brooks, A. E. Gelman, G. L. Jones, and X. L. Meng. Chapman & Hall/CRC, Boca Raton, pp. 3–48.
Markov processes
Markov models
Stochastic processes
Stochastic models
Probability theorems
Asymptotic theory (statistics)
Normal distribution