HOME

TheInfoList



OR:

In
probability theory Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
, the Vysochanskij– Petunin inequality gives a lower bound for the
probability Probability is a branch of mathematics and statistics concerning events and numerical descriptions of how likely they are to occur. The probability of an event is a number between 0 and 1; the larger the probability, the more likely an e ...
that a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
with finite
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
lies within a certain number of
standard deviation In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
s of the variable's
mean A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
, or equivalently an upper bound for the probability that it lies further away. The sole restrictions on the
distribution Distribution may refer to: Mathematics *Distribution (mathematics), generalized functions used to formulate solutions of partial differential equations *Probability distribution, the probability of a particular value or value range of a varia ...
are that it be
unimodal In mathematics, unimodality means possessing a unique mode. More generally, unimodality means there is only a single highest value, somehow defined, of some mathematical object. Unimodal probability distribution In statistics, a unimodal p ...
and have finite
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
; here ''unimodal'' implies that it is a continuous probability distribution except at the
mode Mode ( meaning "manner, tune, measure, due measure, rhythm, melody") may refer to: Arts and entertainment * MO''D''E (magazine), a defunct U.S. women's fashion magazine * ''Mode'' magazine, a fictional fashion magazine which is the setting fo ...
, which may have a non-zero probability.


Theorem

Let X be a random variable with unimodal distribution, and \alpha\in \mathbb R. If we define \rho=\sqrt then for any r>0, :\begin \operatorname(, X-\alpha, \ge r)\le \begin \frac&r\ge \sqrt\rho \\ \frac-\frac&r\le \sqrt\rho. \\ \end \end


Relation to Gauss's inequality

Taking \alpha equal to a mode of X yields the first case of
Gauss's inequality In probability theory, Gauss's inequality (or the Gauss inequality) gives an upper bound on the probability that a unimodal random variable lies more than any given distance from its mode. Let ''X'' be a unimodal random variable with mode ''m'', a ...
.


Tightness of Bound

Without loss of generality, assume \alpha=0 and \rho=1. * If r<1, the left-hand side can equal one, so the bound is useless. * If r\ge \sqrt, the bound is tight when X=0 with probability 1-\frac and is otherwise distributed uniformly in the interval \left \frac,\frac\right/math>. * If 1\le r\le \sqrt, the bound is tight when X=r with probability \frac-\frac and is otherwise distributed uniformly in the interval \left \frac,r\right/math>.


Specialization to mean and variance

If X has mean \mu and finite, non-zero variance \sigma^2, then taking \alpha=\mu and r=\lambda \sigma gives that for any \lambda > \sqrt = 1.63299..., :\operatorname(\left, X-\mu\\geq \lambda\sigma)\leq\frac.


Proof Sketch

For a relatively elementary proof see.Pukelsheim, F., 1994. The Three Sigma Rule. ''The American Statistician'', 48(2), pp.88-91
/ref> The rough idea behind the proof is that there are two cases: one where the mode of X is close to \alpha compared to r, in which case we can show \operatorname(, X-\alpha, \ge r)\le \frac, and one where the mode of X is far from \alpha compared to r, in which case we can show \operatorname(, X-\alpha, \ge r)\le \frac-\frac. Combining these two cases gives \operatorname(, X-\alpha, \ge r)\le \max\left(\frac,\frac-\frac\right). When \frac=\sqrt, the two cases give the same value.


Properties

The theorem refines
Chebyshev's inequality In probability theory, Chebyshev's inequality (also called the Bienaymé–Chebyshev inequality) provides an upper bound on the probability of deviation of a random variable (with finite variance) from its mean. More specifically, the probability ...
by including the factor of 4/9, made possible by the condition that the distribution be unimodal. It is common, in the construction of
control chart Control charts are graphical plots used in production control to determine whether quality and manufacturing processes are being controlled under stable conditions. (ISO 7870-1) The hourly status is arranged on the graph, and the occurrence of ...
s and other statistical heuristics, to set , corresponding to an upper probability bound of 4/81= 0.04938..., and to construct ''3-sigma'' limits to bound ''nearly all'' (i.e. 95%) of the values of a process output. Without unimodality Chebyshev's inequality would give a looser bound of .


One-sided version

An improved version of the Vysochanskij-Petunin inequality for one-sided tail bounds exists. For a unimodal random variable X with mean \mu and variance \sigma^2 , and r \geq 0, the one-sided Vysochanskij-Petunin inequality (Mercadier and Strobel, 2021) holds as follows: :\mathbb(X-\mu\geq r)\leq \begin \dfrac\dfrac & \mboxr^\geq\dfrac\sigma^2,\\ \dfrac\dfrac-\dfrac & \mbox \end The one-sided Vysochanskij-Petunin inequality, as well as the related Cantelli inequality, can for instance be relevant in the financial area, in the sense of "how bad can losses get."


Proof

The proof is very similar to that of
Cantelli's inequality In probability theory, Cantelli's inequality (also called the Chebyshev-Cantelli inequality and the one-sided Chebyshev inequality) is an improved version of Chebyshev's inequality for one-sided tail bounds. The inequality states that, for \lambda ...
. For any u\ge 0, :\begin \mathbb(X-\mu\geq r)&=\mathbb((X+u)-\mu\geq r+u)\\ &\le \mathbb(, (X+u)-\mu), \geq r+u).\\ \end Then we can apply the Vysochanskij-Petunin inequality. With \rho^2=\mathbb E (X+u)-\mu)^2u^2+\sigma^2, we have: : \begin \mathbb(, (X+u)-\mu), \geq r+u) &\le \begin \frac \frac & r+u\ge \sqrt\rho\\ \frac \frac-\frac & r+u\le \sqrt\rho \end. \end As in the proof of Cantelli's inequality, it can be shown that the minimum of \frac over all u\ge 0 is achieved at u=\sigma^2/r. Plugging in this value of u and simplifying yields the desired inequality.


Generalisation

Dharmadhikari and Joag-Dev generalised the VP inequality to deviations from an arbitrary point and moments of order k other than 2 : \begin P(, X-\alpha, \geq r) \leq \max \left\ \\ \end where : \begin \tau_k=E\left(, X-\alpha, ^k\right), s>(k+1), s(s-k-1)^k=k^k \end The standard form of the inequality can be recovered by setting k = 2 which leads to a unique value of s = 4.


See also

*
Gauss's inequality In probability theory, Gauss's inequality (or the Gauss inequality) gives an upper bound on the probability that a unimodal random variable lies more than any given distance from its mode. Let ''X'' be a unimodal random variable with mode ''m'', a ...
, a similar result for the distance from the mode rather than the mean *
Rule of three (statistics) In statistical analysis, the rule of three states that if a certain event did not occur in a sample with Design of experiments, subjects, the interval from 0 to 3/ is a 95% confidence interval for the rate of occurrences in the population (stat ...
, a similar result for the Bernoulli distribution


References

*
Report (on cancer diagnosis) by Petunin and others stating theorem in English
{{DEFAULTSORT:Vysochanskij-Petunin inequality Probabilistic inequalities Statistical inequalities