In
probability theory
Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
, there exist several different notions of convergence of sequences of random variables, including ''convergence in probability'', ''convergence in distribution'', and ''almost sure convergence''. The different notions of convergence capture different properties about the sequence, with some notions of convergence being stronger than others. For example, convergence in distribution tells us about the limit
distribution of a sequence of random variables. This is a weaker notion than convergence in probability, which tells us about the value a random variable will take, rather than just the distribution.
The concept is important in probability theory, and its applications to
statistics
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
and
stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
es. The same concepts are known in more general
mathematics
Mathematics is a field of study that discovers and organizes methods, Mathematical theory, theories and theorems that are developed and Mathematical proof, proved for the needs of empirical sciences and mathematics itself. There are many ar ...
as stochastic convergence and they formalize the idea that certain properties of a sequence of essentially random or unpredictable events can sometimes be expected to settle down into a behavior that is essentially unchanging when items far enough into the sequence are studied. The different possible notions of convergence relate to how such a behavior can be characterized: two readily understood behaviors are that the sequence eventually takes a constant value, and that values in the sequence continue to change but can be described by an unchanging probability distribution.
Background
"Stochastic convergence" formalizes the idea that a sequence of essentially random or unpredictable events can sometimes be expected to settle into a pattern. The pattern may for instance be
*
Convergence
Convergence may refer to:
Arts and media Literature
*''Convergence'' (book series), edited by Ruth Nanda Anshen
*Convergence (comics), "Convergence" (comics), two separate story lines published by DC Comics:
**A four-part crossover storyline that ...
in the classical sense to a fixed value, perhaps itself coming from a random event
*An increasing similarity of outcomes to what a purely deterministic function would produce
*An increasing preference towards a certain outcome
*An increasing "aversion" against straying far away from a certain outcome
*That the probability distribution describing the next outcome may grow increasingly similar to a certain distribution
Some less obvious, more theoretical patterns could be
*That the series formed by calculating the
expected value
In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
of the outcome's distance from a particular value may converge to 0
*That the variance of the
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
describing the next event grows smaller and smaller.
These other types of patterns that may arise are reflected in the different types of stochastic convergence that have been studied.
While the above discussion has related to the convergence of a single series to a limiting value, the notion of the convergence of two series towards each other is also important, but this is easily handled by studying the sequence defined as either the difference or the ratio of the two series.
For example, if the average of ''n''
independent
Independent or Independents may refer to:
Arts, entertainment, and media Artist groups
* Independents (artist group), a group of modernist painters based in Pennsylvania, United States
* Independentes (English: Independents), a Portuguese artist ...
random variables
, all having the same finite
mean
A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
and
variance
In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
, is given by
:
then as
tends to infinity,
converges ''in probability'' (see below) to the common
mean
A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
,
, of the random variables
. This result is known as the
weak law of large numbers. Other forms of convergence are important in other useful theorems, including the
central limit theorem
In probability theory, the central limit theorem (CLT) states that, under appropriate conditions, the Probability distribution, distribution of a normalized version of the sample mean converges to a Normal distribution#Standard normal distributi ...
.
Throughout the following, we assume that
is a sequence of random variables, and
is a random variable, and all of them are defined on the same
probability space
In probability theory, a probability space or a probability triple (\Omega, \mathcal, P) is a mathematical construct that provides a formal model of a random process or "experiment". For example, one can define a probability space which models ...
.
Convergence in distribution
Loosely, with this mode of convergence, we increasingly expect to see the next outcome in a sequence of random experiments becoming better and better modeled by a given
probability distribution
In probability theory and statistics, a probability distribution is a Function (mathematics), function that gives the probabilities of occurrence of possible events for an Experiment (probability theory), experiment. It is a mathematical descri ...
. More precisely, the distribution of the associated random variable in the sequence becomes arbitrarily close to a specified fixed distribution.
Convergence in distribution is the weakest form of convergence typically discussed, since it is implied by all other types of convergence mentioned in this article. However, convergence in distribution is very frequently used in practice; most often it arises from application of the
central limit theorem
In probability theory, the central limit theorem (CLT) states that, under appropriate conditions, the Probability distribution, distribution of a normalized version of the sample mean converges to a Normal distribution#Standard normal distributi ...
.
Definition
A sequence
of real-valued
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
s, with
cumulative distribution function
In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x.
Ever ...
s
, is said to converge in distribution, or converge weakly, or converge in law to a random variable with
cumulative distribution function
In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x.
Ever ...
if
:
for every number
at which
is
continuous.
The requirement that only the continuity points of
should be considered is essential. For example, if
are distributed
uniformly on intervals
, then this sequence converges in distribution to the
degenerate random variable
. Indeed,
for all
In mathematical logic, a universal quantification is a type of quantifier, a logical constant which is interpreted as "given any", "for all", "for every", or "given an arbitrary element". It expresses that a predicate can be satisfied by e ...
when
, and
for all
when
. However, for this limiting random variable
, even though
for all
. Thus the convergence of cdfs fails at the point
where
is discontinuous.
Convergence in distribution may be denoted as
where
is the law (probability distribution) of . For example, if is standard normal we can write
.
For
random vector
In probability, and statistics, a multivariate random variable or random vector is a list or vector of mathematical variables each of whose value is unknown, either because the value has not yet occurred or because there is imperfect knowledge ...
s
the convergence in distribution is defined similarly. We say that this sequence converges in distribution to a random -vector if
:
for every
which is a
continuity set of .
The definition of convergence in distribution may be extended from random vectors to more general
random elements in arbitrary
metric space
In mathematics, a metric space is a Set (mathematics), set together with a notion of ''distance'' between its Element (mathematics), elements, usually called point (geometry), points. The distance is measured by a function (mathematics), functi ...
s, and even to the “random variables” which are not measurable — a situation which occurs for example in the study of
empirical processes. This is the “weak convergence of laws without laws being defined” — except asymptotically.
In this case the term weak convergence is preferable (see
weak convergence of measures), and we say that a sequence of random elements converges weakly to (denoted as ) if
:
for all continuous bounded functions . Here E* denotes the ''outer expectation'', that is the expectation of a “smallest measurable function that dominates ”.
Properties
* Since
, the convergence in distribution means that the probability for to be in a given range is approximately equal to the probability that the value of is in that range, provided is
sufficiently large
In the mathematical areas of number theory and analysis, an infinite sequence or a function is said to eventually have a certain property, if it does not have the said property across all its ordered instances, but will after some instances have ...
.
*In general, convergence in distribution does not imply that the sequence of corresponding
probability density function
In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a Function (mathematics), function whose value at any given sample (or point) in the sample space (the s ...
s will also converge. As an example one may consider random variables with densities . These random variables converge in distribution to a uniform ''U''(0, 1), whereas their densities do not converge at all.
** However, according to ''Scheffé’s theorem'', convergence of the
probability density function
In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a Function (mathematics), function whose value at any given sample (or point) in the sample space (the s ...
s implies convergence in distribution.
* The
portmanteau lemma provides several equivalent definitions of convergence in distribution. Although these definitions are less intuitive, they are used to prove a number of statistical theorems. The lemma states that converges in distribution to if and only if any of the following statements are true:
**
for all continuity points of
;
**
for all
bounded,
continuous function
In mathematics, a continuous function is a function such that a small variation of the argument induces a small variation of the value of the function. This implies there are no abrupt changes in value, known as '' discontinuities''. More preci ...
s
(where
denotes the
expected value
In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
operator);
**
for all bounded,
Lipschitz function
In mathematical analysis, Lipschitz continuity, named after German mathematician Rudolf Lipschitz, is a strong form of uniform continuity for functions. Intuitively, a Lipschitz continuous function is limited in how fast it can change: there e ...
s
;
**
for all nonnegative, continuous functions
;
**
for every
open set
In mathematics, an open set is a generalization of an Interval (mathematics)#Definitions_and_terminology, open interval in the real line.
In a metric space (a Set (mathematics), set with a metric (mathematics), distance defined between every two ...
;
**
for every
closed set
In geometry, topology, and related branches of mathematics, a closed set is a Set (mathematics), set whose complement (set theory), complement is an open set. In a topological space, a closed set can be defined as a set which contains all its lim ...
;
**
for all
continuity sets
of random variable
;
**
for every
upper semi-continuous
In mathematical analysis, semicontinuity (or semi-continuity) is a property of extended real-valued functions that is weaker than continuity. An extended real-valued function f is upper (respectively, lower) semicontinuous at a point x_0 if, r ...
function
bounded above;
**
for every
lower semi-continuous function
bounded below.
* The
continuous mapping theorem states that for a continuous function , if the sequence converges in distribution to , then converges in distribution to .
** Note however that convergence in distribution of to and to does in general ''not'' imply convergence in distribution of to or of to .
*
Lévy’s continuity theorem: The sequence converges in distribution to if and only if the sequence of corresponding
characteristic function In mathematics, the term "characteristic function" can refer to any of several distinct concepts:
* The indicator function of a subset, that is the function
\mathbf_A\colon X \to \,
which for a given subset ''A'' of ''X'', has value 1 at points ...
s
converges pointwise to the characteristic function of .
* Convergence in distribution is
metrizable
In topology and related areas of mathematics, a metrizable space is a topological space that is homeomorphic to a metric space. That is, a topological space (X, \tau) is said to be metrizable if there is a metric d : X \times X \to , \infty) suc ...
by the
Lévy–Prokhorov metric.
* A natural link to convergence in distribution is the
Skorokhod's representation theorem.
Convergence in probability
The basic idea behind this type of convergence is that the probability of an “unusual” outcome becomes smaller and smaller as the sequence progresses.
The concept of convergence in probability is used very often in statistics. For example, an estimator is called
consistent
In deductive logic, a consistent theory is one that does not lead to a logical contradiction. A theory T is consistent if there is no formula \varphi such that both \varphi and its negation \lnot\varphi are elements of the set of consequences ...
if it converges in probability to the quantity being estimated. Convergence in probability is also the type of convergence established by the
weak law of large numbers.
Definition
A sequence of random variables converges in probability towards the random variable ''X'' if for all ''ε'' > 0
:
More explicitly, let ''P''
''n''(''ε'') be the probability that ''X''
''n'' is outside the ball of radius ''ε'' centered at ''X''. Then is said to converge in probability to ''X'' if for any and any ''δ'' > 0 there exists a number ''N'' (which may depend on ''ε'' and ''δ'') such that for all ''n'' ≥ ''N'', ''P''
''n''(''ε'') < ''δ'' (the definition of limit).
Notice that for the condition to be satisfied, it is not possible that for each ''n'' the random variables ''X'' and ''X''
''n'' are independent (and thus convergence in probability is a condition on the joint cdf's, as opposed to convergence in distribution, which is a condition on the individual cdf's), unless ''X'' is deterministic like for the weak law of large numbers. At the same time, the case of a deterministic ''X'' cannot, whenever the deterministic value is a discontinuity point (not isolated), be handled by convergence in distribution, where discontinuity points have to be explicitly excluded.
Convergence in probability is denoted by adding the letter ''p'' over an arrow indicating convergence, or using the "plim" probability limit operator:
For random elements on a
separable metric space , convergence in probability is defined similarly by
:
Properties
* Convergence in probability implies convergence in distribution.
">roof/sup>
* In the opposite direction, convergence in distribution implies convergence in probability when the limiting random variable ''X'' is a constant. ">roof/sup>
* Convergence in probability does not imply almost sure convergence. ">roof/sup>
* The continuous mapping theorem states that for every continuous function , if , then also
* Convergence in probability defines a topology
Topology (from the Greek language, Greek words , and ) is the branch of mathematics concerned with the properties of a Mathematical object, geometric object that are preserved under Continuous function, continuous Deformation theory, deformat ...
on the space of random variables over a fixed probability space. This topology is metrizable
In topology and related areas of mathematics, a metrizable space is a topological space that is homeomorphic to a metric space. That is, a topological space (X, \tau) is said to be metrizable if there is a metric d : X \times X \to , \infty) suc ...
by the '' Ky Fan metric'': or alternately by this metric
Counterexamples
Not every sequence of random variables which converges to another random variable in distribution also converges in probability to that random variable. As an example, consider a sequence of standard normal random variables and a second sequence . Notice that the distribution of is equal to the distribution of for all , but:
which does not converge to . So we do not have convergence in probability.
Almost sure convergence
This is the type of stochastic convergence that is most similar to pointwise convergence
In mathematics, pointwise convergence is one of Modes of convergence (annotated index), various senses in which a sequence of function (mathematics), functions can Limit (mathematics), converge to a particular function. It is weaker than uniform co ...
known from elementary real analysis
In mathematics, the branch of real analysis studies the behavior of real numbers, sequences and series of real numbers, and real functions. Some particular properties of real-valued sequences and functions that real analysis studies include co ...
.
Definition
To say that the sequence converges almost surely or almost everywhere or with probability 1 or strongly towards ''X'' means that
This means that the values of approach the value of ''X'', in the sense that events for which does not converge to ''X'' have probability 0 (see ''Almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (with respect to the probability measure). In other words, the set of outcomes on which the event does not occur ha ...
''). Using the probability space and the concept of the random variable as a function from Ω to R, this is equivalent to the statement
Using the notion of the limit superior of a sequence of sets, almost sure convergence can also be defined as follows:
Almost sure convergence is often denoted by adding the letters ''a.s.'' over an arrow indicating convergence:
For generic random elements on a metric space
In mathematics, a metric space is a Set (mathematics), set together with a notion of ''distance'' between its Element (mathematics), elements, usually called point (geometry), points. The distance is measured by a function (mathematics), functi ...
, convergence almost surely is defined similarly:
Properties
* Almost sure convergence implies convergence in probability (by Fatou's lemma
In mathematics, Fatou's lemma establishes an inequality (mathematics), inequality relating the Lebesgue integral of the limit superior and limit inferior, limit inferior of a sequence of function (mathematics), functions to the limit inferior of ...
), and hence implies convergence in distribution. It is the notion of convergence used in the strong law of large numbers
In probability theory, the law of large numbers is a mathematical law that states that the average of the results obtained from a large number of independent random samples converges to the true value, if it exists. More formally, the law o ...
.
* The concept of almost sure convergence does not come from a topology
Topology (from the Greek language, Greek words , and ) is the branch of mathematics concerned with the properties of a Mathematical object, geometric object that are preserved under Continuous function, continuous Deformation theory, deformat ...
on the space of random variables. This means there is no topology on the space of random variables such that the almost surely convergent sequences are exactly the converging sequences with respect to that topology. In particular, there is no metric of almost sure convergence.
Counterexamples
Consider a sequence of independent random variables such that and . For we have which converges to hence in probability.
Since and the events are independent, second Borel Cantelli Lemma ensures that hence the sequence does not converge to almost everywhere (in fact the set on which this sequence does not converge to has probability ).
Sure convergence or pointwise convergence
To say that the sequence of random variables
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. The term 'random variable' in its mathematical definition refers ...
(''X''''n'') defined over the same probability space
In probability theory, a probability space or a probability triple (\Omega, \mathcal, P) is a mathematical construct that provides a formal model of a random process or "experiment". For example, one can define a probability space which models ...
(i.e., a random process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Stoc ...
) converges surely or everywhere or pointwise towards ''X'' means
where Ω is the sample space
In probability theory, the sample space (also called sample description space, possibility space, or outcome space) of an experiment or random trial is the set of all possible outcomes or results of that experiment. A sample space is usually den ...
of the underlying probability space
In probability theory, a probability space or a probability triple (\Omega, \mathcal, P) is a mathematical construct that provides a formal model of a random process or "experiment". For example, one can define a probability space which models ...
over which the random variables are defined.
This is the notion of pointwise convergence
In mathematics, pointwise convergence is one of Modes of convergence (annotated index), various senses in which a sequence of function (mathematics), functions can Limit (mathematics), converge to a particular function. It is weaker than uniform co ...
of a sequence of functions extended to a sequence of random variables
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. The term 'random variable' in its mathematical definition refers ...
. (Note that random variables themselves are functions).
Sure convergence of a random variable implies all the other kinds of convergence stated above, but there is no payoff in probability theory
Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
by using sure convergence compared to using almost sure convergence. The difference between the two only exists on sets with probability zero. This is why the concept of sure convergence of random variables is very rarely used.
Convergence in mean
Given a real number , we say that the sequence converges in the ''r''-th mean (or in the ''Lr''-norm) towards the random variable ''X'', if the -th absolute moments (, ''Xn'', ''r '') and (, ''X'', ''r '') of and ''X'' exist, and
:
where the operator E denotes the expected value
In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
. Convergence in -th mean tells us that the expectation of the -th power of the difference between and converges to zero.
This type of convergence is often denoted by adding the letter ''Lr'' over an arrow indicating convergence:
The most important cases of convergence in ''r''-th mean are:
* When converges in ''r''-th mean to ''X'' for ''r'' = 1, we say that converges in mean to ''X''.
* When converges in ''r''-th mean to ''X'' for ''r'' = 2, we say that converges in mean square (or in quadratic mean) to ''X''.
Convergence in the ''r''-th mean, for ''r'' ≥ 1, implies convergence in probability (by Markov's inequality
In probability theory, Markov's inequality gives an upper bound on the probability that a non-negative random variable is greater than or equal to some positive Constant (mathematics), constant. Markov's inequality is tight in the sense that for e ...
). Furthermore, if ''r'' > ''s'' ≥ 1, convergence in ''r''-th mean implies convergence in ''s''-th mean. Hence, convergence in mean square implies convergence in mean.
Additionally,
:
The converse is not necessarily true, however it is true if (by a more general version of Scheffé's lemma).
Properties
Provided the probability space is complete:
* If and , then almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (with respect to the probability measure). In other words, the set of outcomes on which the event does not occur ha ...
.
* If and , then almost surely.
* If and , then almost surely.
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ).
* None of the above statements are true for convergence in distribution.
The chain of implications between the various notions of convergence are noted in their respective sections. They are, using the arrow notation:
:
These properties, together with a number of other special cases, are summarized in the following list:
* Almost sure convergence implies convergence in probability: ">roof/sup>
*:
* Convergence in probability implies there exists a sub-sequence which almost surely converges:
*:
* Convergence in probability implies convergence in distribution: ">roof/sup>
*:
* Convergence in ''r''-th order mean implies convergence in probability:
*:
* Convergence in ''r''-th order mean implies convergence in lower order mean, assuming that both orders are greater than or equal to one:
*: provided ''r'' ≥ ''s'' ≥ 1.
* If ''X''''n'' converges in distribution to a constant ''c'', then ''X''''n'' converges in probability to ''c'': ">roof/sup>
*: provided ''c'' is a constant.
* If converges in distribution to ''X'' and the difference between ''Xn'' and ''Yn'' converges in probability to zero, then ''Yn'' also converges in distribution to ''X'': ">roof/sup>
*:
* If converges in distribution to ''X'' and ''Yn'' converges in distribution to a constant ''c'', then the joint vector converges in distribution to : ">roof/sup>
*: provided ''c'' is a constant.
*:Note that the condition that converges to a constant is important, if it were to converge to a random variable ''Y'' then we wouldn't be able to conclude that converges to .
* If ''Xn'' converges in probability to ''X'' and ''Yn'' converges in probability to ''Y'', then the joint vector converges in probability to : ">roof/sup>
*:
* If converges in probability to ''X'', and if for all ''n'' and some ''b'', then converges in ''r''th mean to ''X'' for all . In other words, if converges in probability to ''X'' and all random variables are almost surely bounded above and below, then converges to ''X'' also in any ''r''th mean.
* Almost sure representation. Usually, convergence in distribution does not imply convergence almost surely. However, for a given sequence which converges in distribution to ''X''0 it is always possible to find a new probability space (Ω, ''F'', P) and random variables defined on it such that ''Yn'' is equal in distribution to for each , and ''Yn'' converges to ''Y''0 almost surely.
* If for all ''ε'' > 0,
*::
*:then we say that ''converges almost completely'', or ''almost in probability'' towards ''X''. When converges almost completely towards ''X'' then it also converges almost surely to ''X''. In other words, if converges in probability to ''X'' sufficiently quickly (i.e. the above sequence of tail probabilities is summable for all ), then also converges almost surely to ''X''. This is a direct implication from the Borel–Cantelli lemma.
* If is a sum of ''n'' real independent random variables:
*::
*:then converges almost surely if and only if converges in probability. The proof can be found in Page 126 (Theorem 5.3.4) of the book by Kai Lai Chung.
*:However, for a sequence of mutually independent random variables, convergence in probability does not imply almost sure convergence.
* The dominated convergence theorem
In measure theory, Lebesgue's dominated convergence theorem gives a mild sufficient condition under which limits and integrals of a sequence of functions can be interchanged. More technically it says that if a sequence of functions is bounded i ...
gives sufficient conditions for almost sure convergence to imply ''L''1-convergence:
}
*A necessary and sufficient condition for ''L''1 convergence is and the sequence (''Xn'') is uniformly integrable.
*If , the followings are equivalent
**,
**,
** is uniformly integrable.
See also
* Proofs of convergence of random variables
* Convergence of measures
* Convergence in measure
* Continuous stochastic process: the question of continuity of a stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
* Big O in probability notation
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem
* Continuous mapping theorem
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Random variables, Convergence of>X_n, ^r= \mathbb X, ^r
The converse is not necessarily true, however it is true if (by a more general version of Scheffé's lemma).
Properties
Provided the probability space is complete:
* If and , then almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (with respect to the probability measure). In other words, the set of outcomes on which the event does not occur ha ...
.
* If and , then almost surely.
* If and , then almost surely.
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ).
* None of the above statements are true for convergence in distribution.
The chain of implications between the various notions of convergence are noted in their respective sections. They are, using the arrow notation:
:
These properties, together with a number of other special cases, are summarized in the following list:
* Almost sure convergence implies convergence in probability: ">roof/sup>
*:
* Convergence in probability implies there exists a sub-sequence which almost surely converges:
*:
* Convergence in probability implies convergence in distribution: ">roof/sup>
*:
* Convergence in ''r''-th order mean implies convergence in probability:
*:
* Convergence in ''r''-th order mean implies convergence in lower order mean, assuming that both orders are greater than or equal to one:
*: provided ''r'' ≥ ''s'' ≥ 1.
* If ''X''''n'' converges in distribution to a constant ''c'', then ''X''''n'' converges in probability to ''c'': ">roof/sup>
*: provided ''c'' is a constant.
* If converges in distribution to ''X'' and the difference between ''Xn'' and ''Yn'' converges in probability to zero, then ''Yn'' also converges in distribution to ''X'': ">roof/sup>
*:
* If converges in distribution to ''X'' and ''Yn'' converges in distribution to a constant ''c'', then the joint vector converges in distribution to : ">roof/sup>
*: provided ''c'' is a constant.
*:Note that the condition that converges to a constant is important, if it were to converge to a random variable ''Y'' then we wouldn't be able to conclude that converges to .
* If ''Xn'' converges in probability to ''X'' and ''Yn'' converges in probability to ''Y'', then the joint vector converges in probability to : ">roof/sup>
*:
* If converges in probability to ''X'', and if for all ''n'' and some ''b'', then converges in ''r''th mean to ''X'' for all . In other words, if converges in probability to ''X'' and all random variables are almost surely bounded above and below, then converges to ''X'' also in any ''r''th mean.
* Almost sure representation. Usually, convergence in distribution does not imply convergence almost surely. However, for a given sequence which converges in distribution to ''X''0 it is always possible to find a new probability space (Ω, ''F'', P) and random variables defined on it such that ''Yn'' is equal in distribution to for each , and ''Yn'' converges to ''Y''0 almost surely.
* If for all ''ε'' > 0,
*::
*:then we say that ''converges almost completely'', or ''almost in probability'' towards ''X''. When converges almost completely towards ''X'' then it also converges almost surely to ''X''. In other words, if converges in probability to ''X'' sufficiently quickly (i.e. the above sequence of tail probabilities is summable for all ), then also converges almost surely to ''X''. This is a direct implication from the Borel–Cantelli lemma.
* If is a sum of ''n'' real independent random variables:
*::
*:then converges almost surely if and only if converges in probability. The proof can be found in Page 126 (Theorem 5.3.4) of the book by Kai Lai Chung.
*:However, for a sequence of mutually independent random variables, convergence in probability does not imply almost sure convergence.
* The dominated convergence theorem
In measure theory, Lebesgue's dominated convergence theorem gives a mild sufficient condition under which limits and integrals of a sequence of functions can be interchanged. More technically it says that if a sequence of functions is bounded i ...
gives sufficient conditions for almost sure convergence to imply ''L''1-convergence:
}
*A necessary and sufficient condition for ''L''1 convergence is and the sequence (''Xn'') is uniformly integrable.
*If , the followings are equivalent
**,
**
The converse is not necessarily true, however it is true if (by a more general version of Scheffé's lemma).
Properties
Provided the probability space is complete:
* If and , then almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (with respect to the probability measure). In other words, the set of outcomes on which the event does not occur ha ...
.
* If and , then almost surely.
* If and , then almost surely.
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ).
* None of the above statements are true for convergence in distribution.
The chain of implications between the various notions of convergence are noted in their respective sections. They are, using the arrow notation:
:
These properties, together with a number of other special cases, are summarized in the following list:
* Almost sure convergence implies convergence in probability: ">roof/sup>
*:
* Convergence in probability implies there exists a sub-sequence which almost surely converges:
*:
* Convergence in probability implies convergence in distribution: ">roof/sup>
*:
* Convergence in ''r''-th order mean implies convergence in probability:
*:
* Convergence in ''r''-th order mean implies convergence in lower order mean, assuming that both orders are greater than or equal to one:
*: provided ''r'' ≥ ''s'' ≥ 1.
* If ''X''''n'' converges in distribution to a constant ''c'', then ''X''''n'' converges in probability to ''c'': ">roof/sup>
*: provided ''c'' is a constant.
* If converges in distribution to ''X'' and the difference between ''Xn'' and ''Yn'' converges in probability to zero, then ''Yn'' also converges in distribution to ''X'': ">roof/sup>
*:
* If converges in distribution to ''X'' and ''Yn'' converges in distribution to a constant ''c'', then the joint vector converges in distribution to : ">roof/sup>
*: provided ''c'' is a constant.
*:Note that the condition that converges to a constant is important, if it were to converge to a random variable ''Y'' then we wouldn't be able to conclude that converges to .
* If ''Xn'' converges in probability to ''X'' and ''Yn'' converges in probability to ''Y'', then the joint vector converges in probability to : ">roof/sup>
*:
* If converges in probability to ''X'', and if for all ''n'' and some ''b'', then converges in ''r''th mean to ''X'' for all . In other words, if converges in probability to ''X'' and all random variables are almost surely bounded above and below, then converges to ''X'' also in any ''r''th mean.
* Almost sure representation. Usually, convergence in distribution does not imply convergence almost surely. However, for a given sequence which converges in distribution to ''X''0 it is always possible to find a new probability space (Ω, ''F'', P) and random variables defined on it such that ''Yn'' is equal in distribution to for each , and ''Yn'' converges to ''Y''0 almost surely.
* If for all ''ε'' > 0,
*::
*:then we say that ''converges almost completely'', or ''almost in probability'' towards ''X''. When converges almost completely towards ''X'' then it also converges almost surely to ''X''. In other words, if converges in probability to ''X'' sufficiently quickly (i.e. the above sequence of tail probabilities is summable for all ), then also converges almost surely to ''X''. This is a direct implication from the Borel–Cantelli lemma.
* If is a sum of ''n'' real independent random variables:
*::
*:then converges almost surely if and only if converges in probability. The proof can be found in Page 126 (Theorem 5.3.4) of the book by Kai Lai Chung.
*:However, for a sequence of mutually independent random variables, convergence in probability does not imply almost sure convergence.
* The dominated convergence theorem
In measure theory, Lebesgue's dominated convergence theorem gives a mild sufficient condition under which limits and integrals of a sequence of functions can be interchanged. More technically it says that if a sequence of functions is bounded i ...
gives sufficient conditions for almost sure convergence to imply ''L''1-convergence:
}
*A necessary and sufficient condition for ''L''1 convergence is and the sequence (''Xn'') is uniformly integrable.
*If , the followings are equivalent
**,
**,
** is uniformly integrable.
See also
* Proofs of convergence of random variables
* Convergence of measures
* Convergence in measure
* Continuous stochastic process: the question of continuity of a stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
* Big O in probability notation
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem
* Continuous mapping theorem
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Random variables, Convergence of>X_n, ^r\rightarrow \mathbb X, ^r< \infty ,
** is uniformly integrable.
See also
* Proofs of convergence of random variables
* Convergence of measures
* Convergence in measure
* Continuous stochastic process: the question of continuity of a stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
* Big O in probability notation
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem
* Continuous mapping theorem
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Random variables, Convergence of>X_n, ^r\rightarrow \mathbb X, ^r< \infty ,
** is uniformly integrable.
See also
* Proofs of convergence of random variables
* Convergence of measures
* Convergence in measure
* Continuous stochastic process: the question of continuity of a stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
* Big O in probability notation
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem
* Continuous mapping theorem
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Random variables, Convergence of>X_n, ^r= \mathbb X, ^r
The converse is not necessarily true, however it is true if (by a more general version of Scheffé's lemma).
Properties
Provided the probability space is complete:
* If and , then almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (with respect to the probability measure). In other words, the set of outcomes on which the event does not occur ha ...
.
* If and , then almost surely.
* If and , then almost surely.
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ).
* None of the above statements are true for convergence in distribution.
The chain of implications between the various notions of convergence are noted in their respective sections. They are, using the arrow notation:
:
These properties, together with a number of other special cases, are summarized in the following list:
* Almost sure convergence implies convergence in probability: ">roof/sup>
*:
* Convergence in probability implies there exists a sub-sequence which almost surely converges:
*:
* Convergence in probability implies convergence in distribution: ">roof/sup>
*:
* Convergence in ''r''-th order mean implies convergence in probability:
*:
* Convergence in ''r''-th order mean implies convergence in lower order mean, assuming that both orders are greater than or equal to one:
*: provided ''r'' ≥ ''s'' ≥ 1.
* If ''X''''n'' converges in distribution to a constant ''c'', then ''X''''n'' converges in probability to ''c'': ">roof/sup>
*: provided ''c'' is a constant.
* If converges in distribution to ''X'' and the difference between ''Xn'' and ''Yn'' converges in probability to zero, then ''Yn'' also converges in distribution to ''X'': ">roof/sup>
*:
* If converges in distribution to ''X'' and ''Yn'' converges in distribution to a constant ''c'', then the joint vector converges in distribution to : ">roof/sup>
*: provided ''c'' is a constant.
*:Note that the condition that converges to a constant is important, if it were to converge to a random variable ''Y'' then we wouldn't be able to conclude that converges to .
* If ''Xn'' converges in probability to ''X'' and ''Yn'' converges in probability to ''Y'', then the joint vector converges in probability to : ">roof/sup>
*:
* If converges in probability to ''X'', and if for all ''n'' and some ''b'', then converges in ''r''th mean to ''X'' for all . In other words, if converges in probability to ''X'' and all random variables are almost surely bounded above and below, then converges to ''X'' also in any ''r''th mean.
* Almost sure representation. Usually, convergence in distribution does not imply convergence almost surely. However, for a given sequence which converges in distribution to ''X''0 it is always possible to find a new probability space (Ω, ''F'', P) and random variables defined on it such that ''Yn'' is equal in distribution to for each , and ''Yn'' converges to ''Y''0 almost surely.
* If for all ''ε'' > 0,
*::
*:then we say that ''converges almost completely'', or ''almost in probability'' towards ''X''. When converges almost completely towards ''X'' then it also converges almost surely to ''X''. In other words, if converges in probability to ''X'' sufficiently quickly (i.e. the above sequence of tail probabilities is summable for all ), then also converges almost surely to ''X''. This is a direct implication from the Borel–Cantelli lemma.
* If is a sum of ''n'' real independent random variables:
*::
*:then converges almost surely if and only if converges in probability. The proof can be found in Page 126 (Theorem 5.3.4) of the book by Kai Lai Chung.
*:However, for a sequence of mutually independent random variables, convergence in probability does not imply almost sure convergence.
* The dominated convergence theorem
In measure theory, Lebesgue's dominated convergence theorem gives a mild sufficient condition under which limits and integrals of a sequence of functions can be interchanged. More technically it says that if a sequence of functions is bounded i ...
gives sufficient conditions for almost sure convergence to imply ''L''1-convergence:
}
*A necessary and sufficient condition for ''L''1 convergence is and the sequence (''Xn'') is uniformly integrable.
*If , the followings are equivalent
**,
**
The converse is not necessarily true, however it is true if (by a more general version of Scheffé's lemma).
Properties
Provided the probability space is complete:
* If and , then almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (with respect to the probability measure). In other words, the set of outcomes on which the event does not occur ha ...
.
* If and , then almost surely.
* If and , then almost surely.
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ).
* None of the above statements are true for convergence in distribution.
The chain of implications between the various notions of convergence are noted in their respective sections. They are, using the arrow notation:
:
These properties, together with a number of other special cases, are summarized in the following list:
* Almost sure convergence implies convergence in probability: ">roof/sup>
*:
* Convergence in probability implies there exists a sub-sequence which almost surely converges:
*:
* Convergence in probability implies convergence in distribution: ">roof/sup>
*:
* Convergence in ''r''-th order mean implies convergence in probability:
*:
* Convergence in ''r''-th order mean implies convergence in lower order mean, assuming that both orders are greater than or equal to one:
*: provided ''r'' ≥ ''s'' ≥ 1.
* If ''X''''n'' converges in distribution to a constant ''c'', then ''X''''n'' converges in probability to ''c'': ">roof/sup>
*: provided ''c'' is a constant.
* If converges in distribution to ''X'' and the difference between ''Xn'' and ''Yn'' converges in probability to zero, then ''Yn'' also converges in distribution to ''X'': ">roof/sup>
*:
* If converges in distribution to ''X'' and ''Yn'' converges in distribution to a constant ''c'', then the joint vector converges in distribution to : ">roof/sup>
*: provided ''c'' is a constant.
*:Note that the condition that converges to a constant is important, if it were to converge to a random variable ''Y'' then we wouldn't be able to conclude that converges to .
* If ''Xn'' converges in probability to ''X'' and ''Yn'' converges in probability to ''Y'', then the joint vector converges in probability to : ">roof/sup>
*:
* If converges in probability to ''X'', and if for all ''n'' and some ''b'', then converges in ''r''th mean to ''X'' for all . In other words, if converges in probability to ''X'' and all random variables are almost surely bounded above and below, then converges to ''X'' also in any ''r''th mean.
* Almost sure representation. Usually, convergence in distribution does not imply convergence almost surely. However, for a given sequence which converges in distribution to ''X''0 it is always possible to find a new probability space (Ω, ''F'', P) and random variables defined on it such that ''Yn'' is equal in distribution to for each , and ''Yn'' converges to ''Y''0 almost surely.
* If for all ''ε'' > 0,
*::
*:then we say that ''converges almost completely'', or ''almost in probability'' towards ''X''. When converges almost completely towards ''X'' then it also converges almost surely to ''X''. In other words, if converges in probability to ''X'' sufficiently quickly (i.e. the above sequence of tail probabilities is summable for all ), then also converges almost surely to ''X''. This is a direct implication from the Borel–Cantelli lemma.
* If is a sum of ''n'' real independent random variables:
*::
*:then converges almost surely if and only if converges in probability. The proof can be found in Page 126 (Theorem 5.3.4) of the book by Kai Lai Chung.
*:However, for a sequence of mutually independent random variables, convergence in probability does not imply almost sure convergence.
* The dominated convergence theorem
In measure theory, Lebesgue's dominated convergence theorem gives a mild sufficient condition under which limits and integrals of a sequence of functions can be interchanged. More technically it says that if a sequence of functions is bounded i ...
gives sufficient conditions for almost sure convergence to imply ''L''1-convergence:
}
*A necessary and sufficient condition for ''L''1 convergence is and the sequence (''Xn'') is uniformly integrable.
*If , the followings are equivalent
**,
**,
** is uniformly integrable.
See also
* Proofs of convergence of random variables
* Convergence of measures
* Convergence in measure
* Continuous stochastic process: the question of continuity of a stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
* Big O in probability notation
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem
* Continuous mapping theorem
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Random variables, Convergence of>X_n, ^r= \mathbb X, ^r
The converse is not necessarily true, however it is true if (by a more general version of Scheffé's lemma).
Properties
Provided the probability space is complete:
* If and , then almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (with respect to the probability measure). In other words, the set of outcomes on which the event does not occur ha ...
.
* If and , then almost surely.
* If and , then almost surely.
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ).
* None of the above statements are true for convergence in distribution.
The chain of implications between the various notions of convergence are noted in their respective sections. They are, using the arrow notation:
:
These properties, together with a number of other special cases, are summarized in the following list:
* Almost sure convergence implies convergence in probability: ">roof/sup>
*:
* Convergence in probability implies there exists a sub-sequence which almost surely converges:
*:
* Convergence in probability implies convergence in distribution: ">roof/sup>
*:
* Convergence in ''r''-th order mean implies convergence in probability:
*:
* Convergence in ''r''-th order mean implies convergence in lower order mean, assuming that both orders are greater than or equal to one:
*: provided ''r'' ≥ ''s'' ≥ 1.
* If ''X''''n'' converges in distribution to a constant ''c'', then ''X''''n'' converges in probability to ''c'': ">roof/sup>
*: provided ''c'' is a constant.
* If converges in distribution to ''X'' and the difference between ''Xn'' and ''Yn'' converges in probability to zero, then ''Yn'' also converges in distribution to ''X'': ">roof/sup>
*:
* If converges in distribution to ''X'' and ''Yn'' converges in distribution to a constant ''c'', then the joint vector converges in distribution to : ">roof/sup>
*: provided ''c'' is a constant.
*:Note that the condition that converges to a constant is important, if it were to converge to a random variable ''Y'' then we wouldn't be able to conclude that converges to .
* If ''Xn'' converges in probability to ''X'' and ''Yn'' converges in probability to ''Y'', then the joint vector converges in probability to : ">roof/sup>
*:
* If converges in probability to ''X'', and if for all ''n'' and some ''b'', then converges in ''r''th mean to ''X'' for all . In other words, if converges in probability to ''X'' and all random variables are almost surely bounded above and below, then converges to ''X'' also in any ''r''th mean.
* Almost sure representation. Usually, convergence in distribution does not imply convergence almost surely. However, for a given sequence which converges in distribution to ''X''0 it is always possible to find a new probability space (Ω, ''F'', P) and random variables defined on it such that ''Yn'' is equal in distribution to for each , and ''Yn'' converges to ''Y''0 almost surely.
* If for all ''ε'' > 0,
*::
*:then we say that ''converges almost completely'', or ''almost in probability'' towards ''X''. When converges almost completely towards ''X'' then it also converges almost surely to ''X''. In other words, if converges in probability to ''X'' sufficiently quickly (i.e. the above sequence of tail probabilities is summable for all ), then also converges almost surely to ''X''. This is a direct implication from the Borel–Cantelli lemma.
* If is a sum of ''n'' real independent random variables:
*::
*:then converges almost surely if and only if converges in probability. The proof can be found in Page 126 (Theorem 5.3.4) of the book by Kai Lai Chung.
*:However, for a sequence of mutually independent random variables, convergence in probability does not imply almost sure convergence.
* The dominated convergence theorem
In measure theory, Lebesgue's dominated convergence theorem gives a mild sufficient condition under which limits and integrals of a sequence of functions can be interchanged. More technically it says that if a sequence of functions is bounded i ...
gives sufficient conditions for almost sure convergence to imply ''L''1-convergence:
}
*A necessary and sufficient condition for ''L''1 convergence is and the sequence (''Xn'') is uniformly integrable.
*If , the followings are equivalent
**,
**
The converse is not necessarily true, however it is true if (by a more general version of Scheffé's lemma).
Properties
Provided the probability space is complete:
* If and , then almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (with respect to the probability measure). In other words, the set of outcomes on which the event does not occur ha ...
.
* If and , then almost surely.
* If and , then almost surely.
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ).
* None of the above statements are true for convergence in distribution.
The chain of implications between the various notions of convergence are noted in their respective sections. They are, using the arrow notation:
:
These properties, together with a number of other special cases, are summarized in the following list:
* Almost sure convergence implies convergence in probability: ">roof/sup>
*:
* Convergence in probability implies there exists a sub-sequence which almost surely converges:
*:
* Convergence in probability implies convergence in distribution: ">roof/sup>
*:
* Convergence in ''r''-th order mean implies convergence in probability:
*:
* Convergence in ''r''-th order mean implies convergence in lower order mean, assuming that both orders are greater than or equal to one:
*: provided ''r'' ≥ ''s'' ≥ 1.
* If ''X''''n'' converges in distribution to a constant ''c'', then ''X''''n'' converges in probability to ''c'': ">roof/sup>
*: provided ''c'' is a constant.
* If converges in distribution to ''X'' and the difference between ''Xn'' and ''Yn'' converges in probability to zero, then ''Yn'' also converges in distribution to ''X'': ">roof/sup>
*:
* If converges in distribution to ''X'' and ''Yn'' converges in distribution to a constant ''c'', then the joint vector converges in distribution to : ">roof/sup>
*: provided ''c'' is a constant.
*:Note that the condition that converges to a constant is important, if it were to converge to a random variable ''Y'' then we wouldn't be able to conclude that converges to .
* If ''Xn'' converges in probability to ''X'' and ''Yn'' converges in probability to ''Y'', then the joint vector converges in probability to : ">roof/sup>
*:
* If converges in probability to ''X'', and if for all ''n'' and some ''b'', then converges in ''r''th mean to ''X'' for all . In other words, if converges in probability to ''X'' and all random variables are almost surely bounded above and below, then converges to ''X'' also in any ''r''th mean.
* Almost sure representation. Usually, convergence in distribution does not imply convergence almost surely. However, for a given sequence which converges in distribution to ''X''0 it is always possible to find a new probability space (Ω, ''F'', P) and random variables defined on it such that ''Yn'' is equal in distribution to for each , and ''Yn'' converges to ''Y''0 almost surely.
* If for all ''ε'' > 0,
*::
*:then we say that ''converges almost completely'', or ''almost in probability'' towards ''X''. When converges almost completely towards ''X'' then it also converges almost surely to ''X''. In other words, if converges in probability to ''X'' sufficiently quickly (i.e. the above sequence of tail probabilities is summable for all ), then also converges almost surely to ''X''. This is a direct implication from the Borel–Cantelli lemma.
* If is a sum of ''n'' real independent random variables:
*::
*:then converges almost surely if and only if converges in probability. The proof can be found in Page 126 (Theorem 5.3.4) of the book by Kai Lai Chung.
*:However, for a sequence of mutually independent random variables, convergence in probability does not imply almost sure convergence.
* The dominated convergence theorem
In measure theory, Lebesgue's dominated convergence theorem gives a mild sufficient condition under which limits and integrals of a sequence of functions can be interchanged. More technically it says that if a sequence of functions is bounded i ...
gives sufficient conditions for almost sure convergence to imply ''L''1-convergence:
}
*A necessary and sufficient condition for ''L''1 convergence is and the sequence (''Xn'') is uniformly integrable.
*If , the followings are equivalent
**,
**,
** is uniformly integrable.
See also
* Proofs of convergence of random variables
* Convergence of measures
* Convergence in measure
* Continuous stochastic process: the question of continuity of a stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
* Big O in probability notation
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem
* Continuous mapping theorem
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Random variables, Convergence of>X_n, ^r\rightarrow \mathbb X, ^r< \infty ,
** is uniformly integrable.
See also
* Proofs of convergence of random variables
* Convergence of measures
* Convergence in measure
* Continuous stochastic process: the question of continuity of a stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
* Big O in probability notation
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem
* Continuous mapping theorem
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Random variables, Convergence of>X_n, ^r\rightarrow \mathbb X, ^r< \infty ,
** is uniformly integrable.
See also
* Proofs of convergence of random variables
* Convergence of measures
* Convergence in measure
* Continuous stochastic process: the question of continuity of a stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
* Big O in probability notation
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem
* Continuous mapping theorem
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Random variables, Convergence of>X_n, ^r\rightarrow \mathbb X, ^r< \infty ,
** is uniformly integrable.
See also
* Proofs of convergence of random variables
* Convergence of measures
* Convergence in measure
* Continuous stochastic process: the question of continuity of a stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
* Big O in probability notation
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem
* Continuous mapping theorem
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Convergence (mathematics), Random variables, Convergence of