Law Of Iterated Expectation
   HOME

TheInfoList



OR:

The proposition in
probability theory Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
known as the law of total expectation, the law of iterated expectations (LIE), Adam's law, the tower rule, and the smoothing property of conditional expectation, among other names, states that if X is a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
whose expected value \operatorname(X) is defined, and Y is any random variable on the same
probability space In probability theory, a probability space or a probability triple (\Omega, \mathcal, P) is a mathematical construct that provides a formal model of a random process or "experiment". For example, one can define a probability space which models ...
, then :\operatorname (X) = \operatorname ( \operatorname ( X \mid Y)), i.e., the
expected value In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
of the
conditional expected value In probability theory, the conditional expectation, conditional expected value, or conditional mean of a random variable is its expected value evaluated with respect to the conditional probability distribution. If the random variable can take on ...
of X given Y is the same as the expected value of X. The
conditional expected value In probability theory, the conditional expectation, conditional expected value, or conditional mean of a random variable is its expected value evaluated with respect to the conditional probability distribution. If the random variable can take on ...
\operatorname( X \mid Y ), with Y a random variable, is not a simple number; it is a random variable whose value depends on the value of Y. That is, the conditional expected value of X given the ''event'' Y = y is a number and it is a function of y. If we write g(y) for the value of \operatorname ( X \mid Y = y) then the random variable \operatorname( X \mid Y ) is g( Y ) . One special case states that if is a finite or
countable In mathematics, a Set (mathematics), set is countable if either it is finite set, finite or it can be made in one to one correspondence with the set of natural numbers. Equivalently, a set is ''countable'' if there exists an injective function fro ...
partition of the
sample space In probability theory, the sample space (also called sample description space, possibility space, or outcome space) of an experiment or random trial is the set of all possible outcomes or results of that experiment. A sample space is usually den ...
, then :\operatorname (X) = \sum_i.


Example

Suppose that only two factories supply
light bulb Electric light is an artificial light source powered by electricity. Electric Light may also refer to: * Light fixture, a decorative enclosure for an electric light source * ''Electric Light'' (album), a 2018 album by James Bay * Electric Light ( ...
s to the market. Factory X's bulbs work for an average of 5000 hours, whereas factory Y's bulbs work for an average of 4000 hours. It is known that factory X supplies 60% of the total bulbs available. What is the expected length of time that a purchased bulb will work for? Applying the law of total expectation, we have: : \begin \operatorname (L) &= \operatorname(L \mid X) \operatorname(X)+\operatorname(L \mid Y) \operatorname(Y) \\ pt&= 5000(0.6)+4000(0.4)\\ pt&=4600 \end where * \operatorname (L) is the expected life of the bulb; * \operatorname(X)= is the probability that the purchased bulb was manufactured by factory X; * \operatorname(Y)= is the probability that the purchased bulb was manufactured by factory Y; * \operatorname(L \mid X)=5000 is the expected lifetime of a bulb manufactured by X; * \operatorname(L \mid Y)=4000 is the expected lifetime of a bulb manufactured by Y. Thus each purchased light bulb has an expected lifetime of 4600 hours.


Informal proof

When a joint
probability density function In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a Function (mathematics), function whose value at any given sample (or point) in the sample space (the s ...
is
well defined In mathematics, a well-defined expression or unambiguous expression is an expression whose definition assigns it a unique interpretation or value. Otherwise, the expression is said to be ''not well defined'', ill defined or ''ambiguous''. A func ...
and the expectations are
integrable In mathematics, integrability is a property of certain dynamical systems. While there are several distinct formal definitions, informally speaking, an integrable system is a dynamical system with sufficiently many conserved quantities, or first ...
, we write for the general case \begin \operatorname E(X) &= \int x \Pr =x~dx \\ \operatorname E(X\mid Y=y) &= \int x \Pr =x\mid Y=y~dx \\ \operatorname E( \operatorname E(X\mid Y)) &= \int \left(\int x \Pr =x\mid Y=y~dx \right) \Pr =y~dy \\ &= \int \int x \Pr = x, Y= y~dx ~dy \\ &= \int x \left( \int \Pr = x, Y = y~dy \right) ~dx \\ &= \int x \Pr = x~dx \\ &= \operatorname E(X)\,.\end A similar derivation works for discrete distributions using summation instead of integration. For the specific case of a partition, give each cell of the partition a unique label and let the random variable ''Y'' be the function of the sample space that assigns a cell's label to each point in that cell.


Proof in the general case

Let (\Omega,\mathcal,\operatorname) be a probability space on which two sub σ-algebras \mathcal_1 \subseteq \mathcal_2 \subseteq \mathcal are defined. For a random variable X on such a space, the smoothing law states that if \operatorname /math> is defined, i.e. \min(\operatorname _+ \operatorname _-<\infty, then : \operatorname \operatorname[X \mid \mathcal_2\mid \mathcal_1">_\mid_\mathcal_2.html" ;"title="\operatorname[X \mid \mathcal_2">\operatorname[X \mid \mathcal_2\mid \mathcal_1= \operatorname[X \mid \mathcal_1]\quad\text. Proof. Since a conditional expectation is a Radon–Nikodym theorem, Radon–Nikodym derivative, verifying the following two properties establishes the smoothing law: * \operatorname \operatorname[X \mid \mathcal_2\mid \mathcal_1">_\mid_\mathcal_2.html" ;"title="\operatorname[X \mid \mathcal_2">\operatorname[X \mid \mathcal_2\mid \mathcal_1\mbox \mathcal_1-measurable * \int_ \operatorname \operatorname[X \mid \mathcal_2\mid \mathcal_1">_\mid_\mathcal_2.html" ;"title="\operatorname[X \mid \mathcal_2">\operatorname[X \mid \mathcal_2\mid \mathcal_1\, d\operatorname = \int_ X \, d\operatorname, for all G_1 \in \mathcal_1. The first of these properties holds by definition of the conditional expectation. To prove the second one, : \begin \min\left(\int_X_+\, d\operatorname, \int_X_-\, d\operatorname \right) &\leq \min\left(\int_\Omega X_+\, d\operatorname, \int_\Omega X_-\, d\operatorname\right)\\[4pt] &=\min(\operatorname _+ \operatorname _- < \infty, \end so the integral \textstyle \int_X\, d\operatorname is defined (not equal \infty - \infty). The second property thus holds since G_1 \in \mathcal_1 \subseteq \mathcal_2 implies : \int_ \operatorname \operatorname[X \mid \mathcal_2\mid \mathcal_1">_\mid_\mathcal_2.html" ;"title="\operatorname[X \mid \mathcal_2">\operatorname[X \mid \mathcal_2\mid \mathcal_1\, d\operatorname = \int_ \operatorname[X \mid \mathcal_2] \, d\operatorname = \int_ X \, d\operatorname. Corollary. In the special case when \mathcal_1 = \ and \mathcal_2 = \sigma(Y), the smoothing law reduces to : \operatorname \operatorname[X \mid Y = \operatorname[X">_\mid_Y.html" ;"title="\operatorname[X \mid Y">\operatorname[X \mid Y = \operatorname[X Alternative proof for \operatorname \operatorname[X \mid Y = \operatorname[X">_\mid_Y.html" ;"title="\operatorname[X \mid Y">\operatorname[X \mid Y = \operatorname[X This is a simple consequence of the measure-theoretic definition of conditional expectation. By definition, \operatorname[X \mid Y] := \operatorname[X \mid \sigma(Y)] is a \sigma(Y)-measurable random variable that satisfies : \int_A \operatorname[X \mid Y] \, d\operatorname = \int_A X \, d\operatorname, for every measurable set A \in \sigma(Y) . Taking A = \Omega proves the claim.


See also

* The
fundamental theorem of poker The fundamental theorem of poker is a principle first articulated by David Sklansky that he believes expresses the essential nature of poker as a game of decision-making in the face of incomplete information. The fundamental theorem is stated ...
for one practical application. *
Law of total probability In probability theory, the law (or formula) of total probability is a fundamental rule relating marginal probabilities to conditional probabilities. It expresses the total probability of an outcome which can be realized via several distinct ev ...
*
Law of total variance The law of total variance is a fundamental result in probability theory that expresses the variance of a random variable in terms of its conditional variances and conditional means given another random variable . Informally, it states that the o ...
*
Law of total covariance In probability theory, the law of total covariance, covariance decomposition formula, or conditional covariance formula states that if ''X'', ''Y'', and ''Z'' are random variables on the same probability space, and the covariance of ''X'' and ''Y'' ...
*
Law of total cumulance In probability theory and mathematical statistics, the law of total cumulance is a generalization to cumulants of the law of total probability, the law of total expectation, and the law of total variance. It has applications in the analysis of ti ...
* Product distribution#expectation (application of the Law for proving that the product expectation is the product of expectations)


References

* (Theorem 34.4) *
Christopher Sims Christopher Albert Sims (born October 21, 1942) is an American econometrician and macroeconomist. He is currently the John J.F. Sherrerd '52 University Professor of Economics at Princeton University. Together with Thomas Sargent, he won the N ...

"Notes on Random Variables, Expectations, Probability Densities, and Martingales"
especially equations (16) through (18) {{DEFAULTSORT:Law Of Total Expectation Algebra of random variables Theory of probability distributions Statistical laws