HOME

TheInfoList



OR:

In
probability theory Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
and
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, the continuous uniform distributions or rectangular distributions are a family of symmetric probability distributions. Such a distribution describes an experiment where there is an arbitrary outcome that lies between certain bounds. The bounds are defined by the parameters, a and b, which are the minimum and maximum values. The interval can either be closed (i.e. ,b/math>) or
open Open or OPEN may refer to: Music * Open (band), Australian pop/rock band * The Open (band), English indie rock band * ''Open'' (Blues Image album), 1969 * ''Open'' (Gerd Dudek, Buschi Niebergall, and Edward Vesala album), 1979 * ''Open'' (Go ...
(i.e. (a,b)). Therefore, the distribution is often abbreviated U(a,b), where U stands for uniform distribution. The difference between the bounds defines the interval length; all intervals of the same length on the distribution's support are equally probable. It is the maximum entropy probability distribution for a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
X under no constraint other than that it is contained in the distribution's support.


Definitions


Probability density function

The
probability density function In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a Function (mathematics), function whose value at any given sample (or point) in the sample space (the s ...
of the continuous uniform distribution is f(x) = \begin \dfrac & \text a \le x \le b, \\ pt 0 & \text x < a \ \text \ x > b. \end The values of f(x) at the two boundaries a and b are usually unimportant, because they do not alter the value of \int_c^d f(x)dx over any interval ,d nor of \int_a^b x f(x) \, dx, nor of any higher moment. Sometimes they are chosen to be zero, and sometimes chosen to be \tfrac . The latter is appropriate in the context of estimation by the method of maximum likelihood. In the context of
Fourier analysis In mathematics, Fourier analysis () is the study of the way general functions may be represented or approximated by sums of simpler trigonometric functions. Fourier analysis grew from the study of Fourier series, and is named after Joseph Fo ...
, one may take the value of f(a) or f(b) to be \tfrac , because then the inverse transform of many
integral transform In mathematics, an integral transform is a type of transform that maps a function from its original function space into another function space via integration, where some of the properties of the original function might be more easily charac ...
s of this uniform function will yield back the function itself, rather than a function which is equal "
almost everywhere In measure theory (a branch of mathematical analysis), a property holds almost everywhere if, in a technical sense, the set for which the property holds takes up nearly all possibilities. The notion of "almost everywhere" is a companion notion to ...
", i.e. except on a set of points with zero measure. Also, it is consistent with the
sign function In mathematics, the sign function or signum function (from '' signum'', Latin for "sign") is a function that has the value , or according to whether the sign of a given real number is positive or negative, or the given number is itself zer ...
, which has no such ambiguity. Any probability density function integrates to 1, so the probability density function of the continuous uniform distribution is graphically portrayed as a rectangle where is the base length and is the height. As the base length increases, the height (the density at any particular value within the distribution boundaries) decreases. In terms of mean \mu and variance \sigma ^2 , the probability density function of the continuous uniform distribution is f(x) = \begin \dfrac & \text - \sigma \sqrt \le x - \mu \le \sigma \sqrt , \\ pt 0 & \text . \end


Cumulative distribution function

The
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ever ...
of the continuous uniform distribution is: F(x) = \begin 0 & \text x < a, \\ pt \frac & \text a \le x \le b, \\ pt 1 & \text x > b. \end Its inverse is: F^ (p) = a + p (b - a) \quad \text 0 < p < 1. In terms of mean \mu and variance \sigma ^2 , the cumulative distribution function of the continuous uniform distribution is: F(x) = \begin 0 & \text x - \mu < - \sigma \sqrt , \\ \frac \left( \frac + 1 \right) & \text - \sigma \sqrt \le x - \mu < \sigma \sqrt , \\ 1 & \text x - \mu \ge \sigma \sqrt ; \end its inverse is: F^ (p) = \sigma \sqrt (2p-1) + \mu \quad \text 0 \le p \le 1.


Example 1. Using the continuous uniform distribution function

For a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
X \sim U(0,23), find \Pr(2 < X < 18): \Pr(2 < X < 18) = (18-2) \cdot \frac = \frac . In a graphical representation of the continuous uniform distribution function (x) \text x the area under the curve within the specified bounds, displaying the probability, is a rectangle. For the specific example above, the base would be and the height would be


Example 2. Using the continuous uniform distribution function (conditional)

For a random variable X \sim U(0,23), find \Pr(X > 12 \mid X > 8): \Pr(X > 12 \mid X > 8) = (23-12) \cdot \frac = \frac. The example above is a
conditional probability In probability theory, conditional probability is a measure of the probability of an Event (probability theory), event occurring, given that another event (by assumption, presumption, assertion or evidence) is already known to have occurred. This ...
case for the continuous uniform distribution: given that is true, what is the probability that Conditional probability changes the sample space, so a new interval length has to be calculated, where b = 23 and a' = 8. The graphical representation would still follow Example 1, where the area under the curve within the specified bounds displays the probability; the base of the rectangle would be and the height would be


Generating functions


Moment-generating function

The moment-generating function of the continuous uniform distribution is: M_X = \operatorname\left e^ \right= \int_a^b e^ \frac = \frac = \frac , from which we may calculate the raw moments m_k : m_1 = \frac , m_2 = \frac , m_k = \frac . For a random variable following the continuous uniform distribution, the
expected value In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
is m_1 = \tfrac , and the
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
is m_2 - m_1 ^2 = \tfrac . For the special case a = -b, the probability density function of the continuous uniform distribution is: f(x) = \begin \frac & \text -b \le x \le b, \\ pt 0 & \text ; \end the moment-generating function reduces to the simple form: M_X = \frac .


Cumulant-generating function

For the n-th cumulant of the continuous uniform distribution on the interval is \tfrac , where B_n is the n-th Bernoulli number.


Standard uniform distribution

The continuous uniform distribution with parameters a = 0 and b = 1, i.e. U(0,1), is called the standard uniform distribution. One interesting property of the standard uniform distribution is that if u_1 has a standard uniform distribution, then so does 1 - u_1 . This property can be used for generating antithetic variates, among other things. In other words, this property is known as the
inversion method Inverse transform sampling (also known as inversion sampling, the inverse probability integral transform, the inverse transformation method, or the Nikolai Smirnov (mathematician), Smirnov transform) is a basic method for pseudo-random number sampl ...
where the continuous standard uniform distribution can be used to generate random numbers for any other continuous distribution. If u_1 is a uniform random number with standard uniform distribution, i.e. with U(0,1), then x = F^ (u_1) generates a random number x from any continuous distribution with the specified
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ever ...
F.


Relationship to other functions

As long as the same conventions are followed at the transition points, the probability density function of the continuous uniform distribution may also be expressed in terms of the Heaviside step function as: f(x) = \frac , or in terms of the rectangle function as: f(x) = \frac \ \operatorname \left( \frac \right) . There is no ambiguity at the transition point of the
sign function In mathematics, the sign function or signum function (from '' signum'', Latin for "sign") is a function that has the value , or according to whether the sign of a given real number is positive or negative, or the given number is itself zer ...
. Using the half-maximum convention at the transition points, the continuous uniform distribution may be expressed in terms of the sign function as: f(x) = \frac .


Properties


Moments

The mean (first raw moment) of the continuous uniform distribution is: \operatorname = \int_a^b x \frac = \frac = \frac . The second raw moment of this distribution is: \operatorname\left ^2\right= \int_a^b x^2 \frac = \frac . In general, the n-th raw moment of this distribution is: \operatorname\left ^n\right= \int_a^b x^n \frac = \frac . The variance (second central moment) of this distribution is: \operatorname = \operatorname\left ^2 \right= \int_a^b \left( x - \frac \right) ^2 \frac = \frac .


Order statistics

Let X_1 , ..., X_n be an i.i.d. sample from U(0,1), and let X_ be the k-th
order statistic In statistics, the ''k''th order statistic of a statistical sample is equal to its ''k''th-smallest value. Together with Ranking (statistics), rank statistics, order statistics are among the most fundamental tools in non-parametric statistics and ...
from this sample. X_ has a beta distribution, with parameters k and The expected value is: \operatorname \left _\right= \frac . This fact is useful when making Q–Q plots. The variance is: \operatorname \left _\right= \frac .


Uniformity

The probability that a continuously uniformly distributed random variable falls within any interval of fixed length is independent of the location of the interval itself (but it is dependent on the interval size ( \ell )), so long as the interval is contained in the distribution's support. Indeed, if X \sim U(a,b) and if ,x+ \ell /math> is a subinterval of ,b/math> with fixed \ell > 0, then: \Pr \big( X \in , x + \ell \big) = \int_x^ \frac = \frac , which is independent of x. This fact motivates the distribution's name.


Uniform distribution on more general sets

The uniform distribution can be generalized to sets more general than intervals. Formally, let S be a
Borel set In mathematics, a Borel set is any subset of a topological space that can be formed from its open sets (or, equivalently, from closed sets) through the operations of countable union, countable intersection, and relative complement. Borel sets ...
of positive, finite
Lebesgue measure In measure theory, a branch of mathematics, the Lebesgue measure, named after French mathematician Henri Lebesgue, is the standard way of assigning a measure to subsets of higher dimensional Euclidean '-spaces. For lower dimensions or , it c ...
\lambda (S), i.e. 0 < \lambda (S) < + \infty . The uniform distribution on S can be specified by defining the probability density function to be zero outside S and constantly equal to \tfrac on S. An interesting special case is when the set S is a simplex. It is possible to obtain a uniform distribution on the standard ''n''-vertex simplex in the following way.take ''n'' independent random variables with the same
exponential distribution In probability theory and statistics, the exponential distribution or negative exponential distribution is the probability distribution of the distance between events in a Poisson point process, i.e., a process in which events occur continuousl ...
; denote them by X1,...,Xn; and let Yi := Xi / (sumi Xi). Then, the vector Y1,...,Yn is uniformly distributed on the simplex.


Related distributions

* If ''X'' has a standard uniform distribution, then by the inverse transform sampling method, ''Y'' = − ''λ''−1 ln(''X'') has an
exponential distribution In probability theory and statistics, the exponential distribution or negative exponential distribution is the probability distribution of the distance between events in a Poisson point process, i.e., a process in which events occur continuousl ...
with (rate) parameter ''λ''. * If ''X'' has a standard uniform distribution, then ''Y'' = ''X''''n'' has a beta distribution with parameters (1/''n'',1). As such, * The Irwin–Hall distribution is the sum of ''n'' i.i.d. ''U''(0,1) distributions. * The Bates distribution is the average of ''n'' i.i.d. ''U''(0,1) distributions. * The standard uniform distribution is a special case of the beta distribution, with parameters (1,1). * The sum of two independent uniform distributions ''U''1(a,b)+''U''2(c,d) yields a trapezoidal distribution, symmetric about its mean, on the support +c,b+d The plateau has width equals to the absolute different of the width of ''U''1 and ''U''2. The width of the sloped parts corresponds to the width of the narrowest uniform distribution. ** If the uniform distributions have the same width w, the result is a triangular distribution, symmetric about its mean, on the support +c,a+c+2w ** The sum of two independent, equally distributed, uniform distributions ''U''1(a,b)+''U''2(a,b) yields a symmetric triangular distribution on the support a,2b * The distance between two i.i.d. uniform random variables , ''U''1(a,b)-''U''2(a,b), also has a triangular distribution, although not symmetric, on the support ,b-a


Statistical inference


Estimation of parameters


Estimation of maximum


= Minimum-variance unbiased estimator

= Given a uniform distribution on ,b/math> with unknown b, the minimum-variance unbiased estimator (UMVUE) for the maximum is: \hat _\text = \frac m = m + \frac , where m is the sample maximum and k is the sample size, sampling without replacement (though this distinction almost surely makes no difference for a continuous distribution). This follows for the same reasons as estimation for the discrete distribution, and can be seen as a very simple case of
maximum spacing estimation In statistics, maximum spacing estimation (MSE or MSP), or maximum product of spacing estimation (MPS), is a method for estimating the parameters of a univariate parametric model, statistical model. The method requires maximization of the geometr ...
. This problem is commonly known as the German tank problem, due to application of maximum estimation to estimates of German tank production during
World War II World War II or the Second World War (1 September 1939 – 2 September 1945) was a World war, global conflict between two coalitions: the Allies of World War II, Allies and the Axis powers. World War II by country, Nearly all of the wo ...
.


= Method of moments estimator

= The method of moments estimator is: \hat _ = 2 \bar , where \bar is the sample mean.


= Maximum likelihood estimator

= The maximum likelihood estimator is: \hat _ = m , where m is the sample maximum, also denoted as m = X_ , the maximum
order statistic In statistics, the ''k''th order statistic of a statistical sample is equal to its ''k''th-smallest value. Together with Ranking (statistics), rank statistics, order statistics are among the most fundamental tools in non-parametric statistics and ...
of the sample.


Estimation of minimum

Given a uniform distribution on ,b/math> with unknown ''a'', the maximum likelihood estimator for ''a'' is: \hat a_=\min\, the sample minimum.


Estimation of midpoint

The midpoint of the distribution, \tfrac , is both the mean and the median of the uniform distribution. Although both the sample mean and the sample median are
unbiased estimator In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In stat ...
s of the midpoint, neither is as efficient as the sample mid-range, i.e. the arithmetic mean of the sample maximum and the sample minimum, which is the UMVU estimator of the midpoint (and also the maximum likelihood estimate).


Confidence interval


For the maximum

Let X_1 , X_2 , X_3 , ..., X_n be a sample from U_ , where L is the maximum value in the population. Then X_ = \max ( X_1 , X_2 , X_3 , ..., X_n ) has the Lebesgue–Borel density f = \frac :Nechval KN, Nechval NA, Vasermanis EK, Makeev VY (2002
Constructing shortest-length confidence intervals
Transport and Telecommunication 3 (1) 95-103
f(t) = n \frac \left( \frac \right) ^ \! = n \frac 1 \! \! 1 _ (t), where 1 \! \! 1 _ is the
indicator function In mathematics, an indicator function or a characteristic function of a subset of a set is a function that maps elements of the subset to one, and all other elements to zero. That is, if is a subset of some set , then the indicator functio ...
of ,L. The confidence interval given before is mathematically incorrect, as \Pr \big( \hat, \hat + \varepsilon \ni \theta \big) \ge 1 - \alpha cannot be solved for \varepsilon without knowledge of \theta. However, one can solve \Pr \big( \hat, \hat (1 + \varepsilon) \ni \theta \big) \ge 1 - \alpha for \varepsilon \ge (1 - \alpha) ^ - 1 for any unknown but valid \theta ; one then chooses the smallest \varepsilon possible satisfying the condition above. Note that the interval length depends upon the random variable \hat .


Occurrence and applications

The probabilities for uniform distribution function are simple to calculate due to the simplicity of the function form. Therefore, there are various applications that this distribution can be used for as shown below: hypothesis testing situations, random sampling cases, finance, etc. Furthermore, generally, experiments of physical origin follow a uniform distribution (e.g. emission of radioactive particles). However, it is important to note that in any application, there is the unchanging assumption that the probability of falling in an interval of fixed length is constant.


Economics example for uniform distribution

In the field of economics, usually
demand In economics, demand is the quantity of a goods, good that consumers are willing and able to purchase at various prices during a given time. In economics "demand" for a commodity is not the same thing as "desire" for it. It refers to both the desi ...
and replenishment may not follow the expected normal distribution. As a result, other distribution models are used to better predict probabilities and trends such as Bernoulli process. But according to Wanke (2008), in the particular case of investigating
lead-time A lead time is the :wikt:latency, latency between the initiation and completion of a process. For example, the lead time between the placement of an order and delivery of new cars by a given manufacturer might be between 2 weeks and 6 months, dep ...
for inventory management at the beginning of the life cycle when a completely new product is being analyzed, the uniform distribution proves to be more useful. In this situation, other distribution may not be viable since there is no existing data on the new product or that the demand history is unavailable so there isn't really an appropriate or known distribution. The uniform distribution would be ideal in this situation since the random variable of lead-time (related to demand) is unknown for the new product but the results are likely to range between a plausible range of two values. The
lead-time A lead time is the :wikt:latency, latency between the initiation and completion of a process. For example, the lead time between the placement of an order and delivery of new cars by a given manufacturer might be between 2 weeks and 6 months, dep ...
would thus represent the random variable. From the uniform distribution model, other factors related to
lead-time A lead time is the :wikt:latency, latency between the initiation and completion of a process. For example, the lead time between the placement of an order and delivery of new cars by a given manufacturer might be between 2 weeks and 6 months, dep ...
were able to be calculated such as cycle service level and shortage per cycle. It was also noted that the uniform distribution was also used due to the simplicity of the calculations.


Sampling from an arbitrary distribution

The uniform distribution is useful for sampling from arbitrary distributions. A general method is the inverse transform sampling method, which uses the
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ever ...
(CDF) of the target random variable. This method is very useful in theoretical work. Since simulations using this method require inverting the CDF of the target variable, alternative methods have been devised for the cases where the CDF is not known in closed form. One such method is
rejection sampling In numerical analysis and computational statistics, rejection sampling is a basic technique used to generate observations from a distribution. It is also commonly called the acceptance-rejection method or "accept-reject algorithm" and is a type o ...
. The
normal distribution In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is f(x) = \frac ...
is an important example where the inverse transform method is not efficient. However, there is an exact method, the Box–Muller transformation, which uses the inverse transform to convert two independent uniform
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
s into two independent normally distributed random variables.


Quantization error

In analog-to-digital conversion, a quantization error occurs. This error is either due to rounding or truncation. When the original signal is much larger than one least significant bit (LSB), the quantization error is not significantly correlated with the signal, and has an approximately uniform distribution. The RMS error therefore follows from the variance of this distribution.


Random variate generation

There are many applications in which it is useful to run simulation experiments. Many
programming language A programming language is a system of notation for writing computer programs. Programming languages are described in terms of their Syntax (programming languages), syntax (form) and semantics (computer science), semantics (meaning), usually def ...
s come with implementations to generate pseudo-random numbers which are effectively distributed according to the standard uniform distribution. On the other hand, the uniformly distributed numbers are often used as the basis for non-uniform random variate generation. If u is a value sampled from the standard uniform distribution, then the value a+(b-a)u follows the uniform distribution parameterized by a and b, as described above.


History

While the historical origins in the conception of uniform distribution are inconclusive, it is speculated that the term "uniform" arose from the concept of equiprobability in dice games (note that the dice games would have
discrete Discrete may refer to: *Discrete particle or quantum in physics, for example in quantum theory * Discrete device, an electronic component with just one circuit element, either passive or active, other than an integrated circuit * Discrete group, ...
and not continuous uniform sample space). Equiprobability was mentioned in
Gerolamo Cardano Gerolamo Cardano (; also Girolamo or Geronimo; ; ; 24 September 1501– 21 September 1576) was an Italian polymath whose interests and proficiencies ranged through those of mathematician, physician, biologist, physicist, chemist, astrologer, as ...
's ''Liber de Ludo Aleae'', a manual written in 16th century and detailed on advanced probability calculus in relation to dice.


See also

* Discrete uniform distribution * Beta distribution * Box–Muller transform * Probability plot * Q–Q plot * Rectangular function * Irwin–Hall distribution — In the degenerate case where n=1, the Irwin-Hall distribution generates a uniform distribution between 0 and 1. * Bates distribution — Similar to the Irwin-Hall distribution, but rescaled for n. Like the Irwin-Hall distribution, in the degenerate case where n=1, the Bates distribution generates a uniform distribution between 0 and 1.


References


Further reading

*


External links


Online calculator of Uniform distribution (continuous)
{{DEFAULTSORT:Uniform Distribution (Continuous) Continuous distributions Location-scale family probability distributions su:Sebaran seragam#Kasus kontinyu