In
statistics, a
statistic
A statistic (singular) or sample statistic is any quantity computed from values in a sample which is considered for a statistical purpose. Statistical purposes include estimating a population parameter, describing a sample, or evaluating a hy ...
is ''sufficient'' with respect to a
statistical model
A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of sample data (and similar data from a larger population). A statistical model represents, often in considerably idealized form, ...
and its associated unknown
parameter
A parameter (), generally, is any characteristic that can help in defining or classifying a particular system (meaning an event, project, object, situation, etc.). That is, a parameter is an element of a system that is useful, or critical, when ...
if "no other statistic that can be calculated from the same
sample provides any additional information as to the value of the parameter".
In particular, a statistic is sufficient for a
family
Family (from la, familia) is a group of people related either by consanguinity (by recognized birth) or affinity (by marriage or other relationship). The purpose of the family is to maintain the well-being of its members and of society. Idea ...
of
probability distribution
In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomeno ...
s if the sample from which it is calculated gives no additional information than the statistic, as to which of those probability distributions is the
sampling distribution
In statistics, a sampling distribution or finite-sample distribution is the probability distribution of a given random-sample-based statistic. If an arbitrarily large number of samples, each involving multiple observations (data points), were se ...
.
A related concept is that of linear sufficiency, which is weaker than ''sufficiency'' but can be applied in some cases where there is no sufficient statistic, although it is restricted to linear estimators. The
Kolmogorov structure function In 1973, Andrey Kolmogorov proposed a non-probabilistic approach to statistics and model selection. Let each datum be a finite binary string and a model be a finite set of binary strings. Consider model classes consisting of models of given maximal ...
deals with individual finite data; the related notion there is the algorithmic sufficient statistic.
The concept is due to
Sir Ronald Fisher
Sir Ronald Aylmer Fisher (17 February 1890 – 29 July 1962) was a British polymath who was active as a mathematician, statistician, biologist, geneticist, and academic. For his work in statistics, he has been described as "a genius who ...
in 1920. Stephen Stigler noted in 1973 that the concept of sufficiency had fallen out of favor in
descriptive statistics
A descriptive statistic (in the count noun sense) is a summary statistic that quantitatively describes or summarizes features from a collection of information, while descriptive statistics (in the mass noun sense) is the process of using and a ...
because of the strong dependence on an assumption of the distributional form (see
Pitman–Koopman–Darmois theorem
In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, including the enabling of the user to calculate ...
below), but remained very important in theoretical work.
Background
Roughly, given a set
of
independent identically distributed
In probability theory and statistics, a collection of random variables is independent and identically distributed if each random variable has the same probability distribution as the others and all are mutually independent. This property is usua ...
data conditioned on an unknown parameter
, a sufficient statistic is a function
whose value contains all the information needed to compute any estimate of the parameter (e.g. a
maximum likelihood
In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed sta ...
estimate). Due to the factorization theorem (
see below), for a sufficient statistic
, the probability density can be written as
. From this factorization, it can easily be seen that the maximum likelihood estimate of
will interact with
only through
. Typically, the sufficient statistic is a simple function of the data, e.g. the sum of all the data points.
More generally, the "unknown parameter" may represent a
vector
Vector most often refers to:
*Euclidean vector, a quantity with a magnitude and a direction
*Vector (epidemiology), an agent that carries and transmits an infectious pathogen into another living organism
Vector may also refer to:
Mathematic ...
of unknown quantities or may represent everything about the model that is unknown or not fully specified. In such a case, the sufficient statistic may be a set of functions, called a ''jointly sufficient statistic''. Typically, there are as many functions as there are parameters. For example, for a
Gaussian distribution
In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is
:
f(x) = \frac e^
The parameter \mu i ...
with unknown
mean
There are several kinds of mean in mathematics, especially in statistics. Each mean serves to summarize a given group of data, often to better understand the overall value ( magnitude and sign) of a given data set.
For a data set, the '' ari ...
and
variance
In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of number ...
, the jointly sufficient statistic, from which maximum likelihood estimates of both parameters can be estimated, consists of two functions, the sum of all data points and the sum of all squared data points (or equivalently, the
sample mean
The sample mean (or "empirical mean") and the sample covariance are statistics computed from a sample of data on one or more random variables.
The sample mean is the average value (or mean value) of a sample of numbers taken from a larger po ...
and
sample variance
In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of number ...
).
The concept is equivalent to the statement that,
conditional
Conditional (if then) may refer to:
* Causal conditional, if X then Y, where X is a cause of Y
* Conditional probability, the probability of an event A given that another event B has occurred
*Conditional proof, in logic: a proof that asserts a ...
on the value of a sufficient statistic for a parameter, the
joint probability distribution
Given two random variables that are defined on the same probability space, the joint probability distribution is the corresponding probability distribution on all possible pairs of outputs. The joint distribution can just as well be considere ...
of the data does not depend on that parameter. Both the statistic and the underlying parameter can be vectors.
Mathematical definition
A statistic ''t'' = ''T''(''X'') is sufficient for underlying parameter ''θ'' precisely if the
conditional probability distribution
In probability theory and statistics, given two jointly distributed random variables X and Y, the conditional probability distribution of Y given X is the probability distribution of Y when X is known to be a particular value; in some cases the c ...
of the data ''X'', given the statistic ''t'' = ''T''(''X''), does not depend on the parameter ''θ''.
Alternatively, one can say the statistic ''T''(''X'') is sufficient for ''θ'' if its
mutual information
In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual dependence between the two variables. More specifically, it quantifies the " amount of information" (in units such as ...
with ''θ'' equals the mutual information between ''X'' and ''θ''. In other words, the
data processing inequality The data processing inequality is an information theoretic concept which states that the information content of a signal cannot be increased via a local physical operation. This can be expressed concisely as 'post-processing cannot increase inform ...
becomes an equality:
:
Example
As an example, the sample mean is sufficient for the mean (''μ'') of a
normal distribution
In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is
:
f(x) = \frac e^
The parameter \mu i ...
with known variance. Once the sample mean is known, no further information about ''μ'' can be obtained from the sample itself. On the other hand, for an arbitrary distribution the
median is not sufficient for the mean: even if the median of the sample is known, knowing the sample itself would provide further information about the population mean. For example, if the observations that are less than the median are only slightly less, but observations exceeding the median exceed it by a large amount, then this would have a bearing on one's inference about the population mean.
Fisher–Neyman factorization theorem
''
Fisher's factorization theorem'' or ''factorization criterion'' provides a convenient characterization of a sufficient statistic. If the
probability density function
In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) c ...
is ƒ
''θ''(''x''), then ''T'' is sufficient for ''θ''
if and only if
In logic and related fields such as mathematics and philosophy, "if and only if" (shortened as "iff") is a biconditional logical connective between statements, where either both statements are true or both are false.
The connective is bi ...
nonnegative functions ''g'' and ''h'' can be found such that
:
i.e. the density ƒ can be factored into a product such that one factor, ''h'', does not depend on ''θ'' and the other factor, which does depend on ''θ'', depends on ''x'' only through ''T''(''x''). A general proof of this was given by Halmos and Savage and the theorem is sometimes referred to as the Halmos-Savage factorization theorem. The proofs below handle special cases, but an alternative general proof along the same lines can be given.
It is easy to see that if ''F''(''t'') is a one-to-one function and ''T'' is a sufficient
statistic, then ''F''(''T'') is a sufficient statistic. In particular we can multiply a
sufficient statistic by a nonzero constant and get another sufficient statistic.
Likelihood principle interpretation
An implication of the theorem is that when using likelihood-based inference, two sets of data yielding the same value for the sufficient statistic ''T''(''X'') will always yield the same inferences about ''θ''. By the factorization criterion, the likelihood's dependence on ''θ'' is only in conjunction with ''T''(''X''). As this is the same in both cases, the dependence on ''θ'' will be the same as well, leading to identical inferences.
Proof
Due to Hogg and Craig.
Let
, denote a random sample from a distribution having the
pdf
Portable Document Format (PDF), standardized as ISO 32000, is a file format developed by Adobe in 1992 to present documents, including text formatting and images, in a manner independent of application software, hardware, and operating systems. ...
''f''(''x'', ''θ'') for ''ι'' < ''θ'' < ''δ''. Let ''Y''
1 = ''u''
1(''X''
1, ''X''
2, ..., ''X''
''n'') be a statistic whose pdf is ''g''
1(''y''
1; ''θ''). What we want to prove is that ''Y''
1 = ''u''
1(''X''
1, ''X''
2, ..., ''X''
''n'') is a sufficient statistic for ''θ'' if and only if, for some function ''H'',
:
First, suppose that
:
We shall make the transformation ''y''
''i'' = ''u''
i(''x''
1, ''x''
2, ..., ''x''
''n''), for ''i'' = 1, ..., ''n'', having inverse functions ''x''
''i'' = ''w''
''i''(''y''
1, ''y''
2, ..., ''y''
''n''), for ''i'' = 1, ..., ''n'', and
Jacobian
In mathematics, a Jacobian, named for Carl Gustav Jacob Jacobi, may refer to:
*Jacobian matrix and determinant
*Jacobian elliptic functions
*Jacobian variety
*Intermediate Jacobian
In mathematics, the intermediate Jacobian of a compact Kähler m ...
. Thus,
:
The left-hand member is the joint pdf ''g''(''y''
1, ''y''
2, ..., ''y''
''n''; θ) of ''Y''
1 = ''u''
1(''X''
1, ..., ''X''
''n''), ..., ''Y''
''n'' = ''u''
''n''(''X''
1, ..., ''X''
''n''). In the right-hand member,
is the pdf of
, so that
is the quotient of
and
; that is, it is the conditional pdf
of
given
.
But
, and thus