HOME

TheInfoList



OR:

BRS-inequality is the short name for Bruss-Robertson-Steele inequality. This inequality gives a convenient upper bound for the expected maximum number of non-negative random variables one can sum up without exceeding a given upper bound s > 0. For example, suppose 100 random variables X_1, X_2,..., X_ are all uniformly distributed on
, 1 The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
/math>, not necessarily independent, and let s= 10, say. Let N , s:= N 00, 10/math> be the maximum number of X_j one can select in \ such that their sum does not exceed s= 10. N 00, 10/math> is a random variable, so what can one say about bounds for its expectation? How would an upper bound for E(N , s behave, if one changes the size n of the sample and keeps s fixed, or alternatively, if one keeps n fixed but varies s? What can one say about E(N , s, if the uniform distribution is replaced by another continuous distribution? In all generality, what can one say if each X_k may have its own continuous distribution function F_k?


General problem

Let X_1, X_2,... be a sequence of non-negative random variables (possibly dependent) that are jointly continuously distributed. For n \in \ and s\in \mathbb^+ let N , s/math> be the maximum number of observations among X_1, X_2, ..., X_n that one can sum up without exceeding s. Now, to obtain N , s/math> one may think of looking at the list of all observations, first select the smallest one, then add the second smallest, then the third and so on, as long as the accumulated sum does not exceed s. Hence N
, n The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
/math> can be defined in terms of the increasing order statistics of X_1, X_2,\cdots, X_n, denoted by X_ \le X_ \le \cdots \le X_,, namely by : \begin N , s= \begin 0 &, ~ X_ > s,\\ \max\ &, . \end \end (1) What is the best possible ''general'' upper bound for E(N , s if one requires only the continuity of the joint distribution of all variables? And then, how to compute this bound?


Identically distributed random variables.

Theorem 1 Let X_1, X_2, \cdots, X_n be identically distributed non-negative random variables with absolutely continuous distribution function F. Then : E(N , s\le n F(t), (2) where t := t(n, s) solves the so-called BRS-equation : n \int_0^t x \,dF(x)\, =\, s. (3) As an example, here are the answers for the questions posed at the beginning. Thus let all X_1, X_2,\cdots, X_n be uniformly distributed on
, 1 The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
/math>. Then F(t) = t on
, 1 The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
/math>, and hence dF(x)/dx = 1 on
, 1 The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
/math>. The BRS-equation becomes : n \int_0^t x dx = n t^2/2 = s. The solution is t =\sqrt, and thus from the inequality (2) : E(N , s \le n\,F(t) = n \sqrt= \sqrt. (4) Since one always has N , s\le n, this bound becomes trivial for s \ge nE(X) = n/2. For the example questions with n=100, s=10 this yields E(N 00, 10 \le \sqrt \approx 44.7. As one sees from (4), doubling the sample size n and keeping s fixed, or vice versa, yield for the uniform distribution in the non-trivial case the same upper bound.


Generalised BRS-inequality

Theorem 2. Let X_1, X_2,\cdots, X_n be positive random variables that are jointly distributed such that X_k has an absolutely continuous distribution function F_k, ~k=1, 2, \cdots,n.. If N , s/math> is defined as before, then :E (N , s\le \sum_^n F_k(t), (5) where t := t(n, s) is the unique solution of the BRS-equation :\sum_^n \int_0^t \,x\,dF_k(x) = s. (6) Clearly, if all random variables X_i, i=1, 2, \cdots, n have the same marginal distribution F, then (6) recaptures (3), and (5) recaptures (2). Again it should be pointed out that no independence hypothesis whatsoever is needed.


Refinements of the BRS-inequality

Depending on the type of the distributions F_k, refinements of Theorem 2 can be of true interest. We just mention one of them. Let A , s/math> be the random set of those variables one can sum up to yield the maximum random number N , s/math>, that is, : \#A , s= N , s/math>, and let S_ denote the sum of these variables. The so-called residual s-S_ is by definition always non-negative, and one has: Theorem 3. Let X_1, X_2,\cdots, X_n be jointly continuously distributed with marginal distribution functions F_k, k=1, 2, \cdots,n, and let t := t(n, s) be the solution of (6). Then :E(N , s\le \left( \sum_^n F_k(t(n, s))\right)-\frac . (7) The improvement in (7) compared with (5) therefore consists of :\frac. The expected residual in the numerator is typically difficult to compute or estimate, but there exist nice exceptions. For example, if all X_k are independent exponential random variables, then the memoryless property implies (if s is exceeded) the distributional symmetry of the residual and the overshoot over s. For fixed s one can then show that :\frac \to 1/2 {\rm~as~} n \to \infty. This improvement fluctuates around 1/2, and the convergence to 1/2, (simulations) seems quick.


Source

The first version of the BRS-inequality (Theorem 1) was proved in Lemma 4.1 of F. Thomas Bruss and James B. Robertson (1991). This paper proves moreover that the upper bound is asymptotically tight if the random variables are independent of each other. The generalisation to arbitrary continuous distributions (Theorem 2) is due to J. Michael Steele (2016). Theorem 3 and other refinements of the BRS-inequality are more recent and proved in Bruss (2021).


Applications

The BRS-inequality is a versatile tool since it applies to many types of problems, and since the computation of the BRS-equation is often not very involved. Also, and in particular, one notes that the maximum number N , s/math> always dominates the maximum number of selections under ''any'' additional constraint, such as e.g. for
online In computer technology and telecommunications, online indicates a state of connectivity, and offline indicates a disconnected state. In modern terminology, this usually refers to an Internet connection, but (especially when expressed as "on lin ...
selections without recall. Examples studied in Steele (2016) and Bruss (2021) touch many applications, including comparisons between i.i.d. sequences and non-i.i.d. sequences, problems of condensing
point process In statistics and probability theory, a point process or point field is a set of a random number of mathematical points randomly located on a mathematical space such as the real line or Euclidean space. Kallenberg, O. (1986). ''Random Measures'', ...
es, “awkward” processes,
selection algorithm In computer science, a selection algorithm is an algorithm for finding the kth smallest value in a collection of ordered values, such as numbers. The value that it finds is called the order statistic. Selection includes as special cases the p ...
s,
knapsack problem The knapsack problem is the following problem in combinatorial optimization: :''Given a set of items, each with a weight and a value, determine which items to include in the collection so that the total weight is less than or equal to a given lim ...
s, Borel-Cantelli-type problems, the Bruss-Duerinckx theorem, and online Tiling strategies.


References

Bruss F. T. and Robertson J. B. (1991)'' ’Wald's Lemma’ for Sums of Order Statistics of i.i.d. Random Variables'', Adv. Appl. Probab., Vol. 23, 612-623. Bruss F. T. and Duerinckx M. (2015), '' Resource dependent branching processes and the envelope of societie'', Ann. of Appl. Probab., Vol. 25 (1), 324-372. Steele J.M. (2016), '' The Bruss-Robertson Inequality: Elaborations, Extensions, and Applications'', Math. Applicanda, Vol. 44, No 1, 3-16. Bruss F. T. (2021),''The BRS-inequality and its applications'', Probab. Surveys, Vol.18, 44-76. Probabilistic inequalities