HOME

TheInfoList



OR:

The sample mean (or "empirical mean") and the sample covariance are
statistic A statistic (singular) or sample statistic is any quantity computed from values in a sample which is considered for a statistical purpose. Statistical purposes include estimating a population parameter, describing a sample, or evaluating a hypo ...
s computed from a
sample Sample or samples may refer to: Base meaning * Sample (statistics), a subset of a population – complete data set * Sample (signal), a digital discrete sample of a continuous analog signal * Sample (material), a specimen or small quantity of s ...
of data on one or more
random variables A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
. The sample mean is the average value (or mean value) of a
sample Sample or samples may refer to: Base meaning * Sample (statistics), a subset of a population – complete data set * Sample (signal), a digital discrete sample of a continuous analog signal * Sample (material), a specimen or small quantity of s ...
of numbers taken from a larger population of numbers, where "population" indicates not number of people but the entirety of relevant data, whether collected or not. A sample of 40 companies' sales from the
Fortune 500 The ''Fortune'' 500 is an annual list compiled and published by ''Fortune'' magazine that ranks 500 of the largest United States corporations by total revenue for their respective fiscal years. The list includes publicly held companies, along ...
might be used for convenience instead of looking at the population, all 500 companies' sales. The sample mean is used as an estimator for the population mean, the average value in the entire population, where the estimate is more likely to be close to the population mean if the sample is large and representative. The reliability of the sample mean is estimated using the standard error, which in turn is calculated using the variance of the sample. If the sample is random, the standard error falls with the size of the sample and the sample mean's distribution approaches the normal distribution as the sample size increases. The term "sample mean" can also be used to refer to a vector of average values when the statistician is looking at the values of several variables in the sample, e.g. the sales, profits, and employees of a sample of Fortune 500 companies. In this case, there is not just a sample variance for each variable but a sample variance-covariance matrix (or simply
covariance matrix In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements of ...
) showing also the relationship between each pair of variables. This would be a 3×3 matrix when 3 variables are being considered. The sample covariance is useful in judging the reliability of the sample means as estimators and is also useful as an estimate of the population covariance matrix. Due to their ease of calculation and other desirable characteristics, the sample mean and sample covariance are widely used in statistics to represent the location and dispersion of the
distribution Distribution may refer to: Mathematics *Distribution (mathematics), generalized functions used to formulate solutions of partial differential equations * Probability distribution, the probability of a particular value or value range of a vari ...
of values in the sample, and to estimate the values for the population.


Definition of the sample mean

The sample mean is the average of the values of a variable in a sample, which is the sum of those values divided by the number of values. Using mathematical notation, if a sample of ''N'' observations on variable ''X'' is taken from the population, the sample mean is: : \bar=\frac\sum_^X_. Under this definition, if the sample (1, 4, 1) is taken from the population (1,1,3,4,0,2,1,0), then the sample mean is \bar = (1+4+1)/3 = 2, as compared to the population mean of \mu = (1+1+3+4+0+2+1+0) /8 = 12/8 = 1.5. Even if a sample is random, it is rarely perfectly representative, and other samples would have other sample means even if the samples were all from the same population. The sample (2, 1, 0), for example, would have a sample mean of 1. If the statistician is interested in ''K'' variables rather than one, each observation having a value for each of those ''K'' variables, the overall sample mean consists of ''K'' sample means for individual variables. Let x_ be the ''i''th independently drawn observation (''i''=1,...,''N'') on the ''j''th random variable (''j''=1,...,''K''). These observations can be arranged into ''N'' column vectors, each with ''K'' entries, with the ''K''×1 column vector giving the ''i''-th observations of all variables being denoted \mathbf_i (''i''=1,...,''N''). The sample mean vector \mathbf is a column vector whose ''j''-th element \bar_ is the average value of the ''N'' observations of the ''j''th variable: : \bar_=\frac \sum_^ x_,\quad j=1,\ldots,K. Thus, the sample mean vector contains the average of the observations for each variable, and is written : \mathbf=\frac\sum_^\mathbf_i = \begin \bar_1 \\ \vdots \\ \bar_j \\ \vdots \\ \bar_K \end


Definition of sample covariance

The sample covariance matrix is a ''K''-by-''K'' matrix \textstyle \mathbf=\left q_\right with entries : q_=\frac\sum_^\left( x_-\bar_j \right) \left( x_-\bar_k \right), where q_ is an estimate of the covariance between the th variable and the th variable of the population underlying the data. In terms of the observation vectors, the sample covariance is :\mathbf = \sum_^N (\mathbf_i.-\mathbf) (\mathbf_i.-\mathbf)^\mathrm, Alternatively, arranging the observation vectors as the columns of a matrix, so that :\mathbf = \begin\mathbf_1 & \mathbf_2 & \dots & \mathbf_N \end, which is a matrix of ''K'' rows and ''N'' columns. Here, the sample covariance matrix can be computed as :\mathbf = \frac( \mathbf - \mathbf \,\mathbf_N^\mathrm ) ( \mathbf - \mathbf \,\mathbf_N^\mathrm )^\mathrm, where \mathbf_N is an ''N'' by vector of ones. If the observations are arranged as rows instead of columns, so \mathbf is now a 1×''K'' row vector and \mathbf=\mathbf^\mathrm is an ''N''×''K'' matrix whose column ''j'' is the vector of ''N'' observations on variable ''j'', then applying transposes in the appropriate places yields :\mathbf = \frac( \mathbf - \mathbf_N \mathbf )^\mathrm ( \mathbf - \mathbf_N \mathbf ). Like covariance matrices for random vector, sample covariance matrices are positive semi-definite. To prove it, note that for any matrix \mathbf the matrix \mathbf^T\mathbf is positive semi-definite. Furthermore, a covariance matrix is positive definite if and only if the rank of the \mathbf_i.-\mathbf vectors is K.


Unbiasedness

The sample mean and the sample covariance matrix are unbiased estimates of the mean and the
covariance matrix In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements of ...
of the random vector \textstyle \mathbf, a row vector whose ''j''th element (''j = 1, ..., K'') is one of the random variables. The sample covariance matrix has \textstyle N-1 in the denominator rather than \textstyle N due to a variant of Bessel's correction: In short, the sample covariance relies on the difference between each observation and the sample mean, but the sample mean is slightly correlated with each observation since it is defined in terms of all observations. If the population mean \operatorname(\mathbf) is known, the analogous unbiased estimate : q_=\frac\sum_^N \left( x_-\operatorname(X_j)\right) \left( x_-\operatorname(X_k)\right), using the population mean, has \textstyle N in the denominator. This is an example of why in probability and statistics it is essential to distinguish between
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
s (upper case letters) and realizations of the random variables (lower case letters). The maximum likelihood estimate of the covariance : q_=\frac\sum_^N \left( x_-\bar_j \right) \left( x_-\bar_k \right) for the Gaussian distribution case has ''N'' in the denominator as well. The ratio of 1/''N'' to 1/(''N'' − 1) approaches 1 for large ''N'', so the maximum likelihood estimate approximately equals the unbiased estimate when the sample is large.


Distribution of the sample mean

For each random variable, the sample mean is a good estimator of the population mean, where a "good" estimator is defined as being efficient and unbiased. Of course the estimator will likely not be the true value of the population mean since different samples drawn from the same distribution will give different sample means and hence different estimates of the true mean. Thus the sample mean is a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
, not a constant, and consequently has its own distribution. For a random sample of ''N'' observations on the ''j''th random variable, the sample mean's distribution itself has mean equal to the population mean E(X_j) and variance equal to \sigma^2_j/N, where \sigma^2_j is the population variance. The arithmetic mean of a population, or population mean, is often denoted ''μ''. The sample mean \bar (the arithmetic mean of a sample of values drawn from the population) makes a good estimator of the population mean, as its expected value is equal to the population mean (that is, it is an unbiased estimator). The sample mean is a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
, not a constant, since its calculated value will randomly differ depending on which members of the population are sampled, and consequently it will have its own distribution. For a random sample of ''n'' independent observations, the expected value of the sample mean is : \operatorname E (\bar) = \mu and the variance of the sample mean is : \operatorname(\bar) = \frac n. If the samples are not independent, but correlated, then special care has to be taken in order to avoid the problem of pseudoreplication. If the population is normally distributed, then the sample mean is normally distributed as follows: : \bar \thicksim N\left\. If the population is not normally distributed, the sample mean is nonetheless approximately normally distributed if ''n'' is large and ''σ''2/''n'' < +∞. This is a consequence of the central limit theorem.


Weighted samples

In a weighted sample, each vector \textstyle \textbf_ (each set of single observations on each of the ''K'' random variables) is assigned a weight \textstyle w_i \geq0. Without loss of generality, assume that the weights are normalized: : \sum_^w_i = 1. (If they are not, divide the weights by their sum). Then the weighted mean vector \textstyle \mathbf is given by : \mathbf=\sum_^N w_i \mathbf_i. and the elements q_ of the weighted covariance matrix \textstyle \mathbf are Mark Galassi, Jim Davies, James Theiler, Brian Gough, Gerard Jungman, Michael Booth, and Fabrice Rossi
GNU Scientific Library - Reference manual, Version 2.6
2021.

/ref> : q_=\frac \sum_^N w_i \left( x_-\bar_j \right) \left( x_-\bar_k \right) . If all weights are the same, \textstyle w_=1/N, the weighted mean and covariance reduce to the (biased) sample mean and covariance mentioned above.


Criticism

The sample mean and sample covariance are not robust statistics, meaning that they are sensitive to outliers. As robustness is often a desired trait, particularly in real-world applications, robust alternatives may prove desirable, notably quantile-based statistics such as the
sample median Sample or samples may refer to: Base meaning * Sample (statistics), a subset of a population – complete data set * Sample (signal), a digital discrete sample of a continuous analog signal * Sample (material), a specimen or small quantity of so ...
for location,The World Question Center 2006: The Sample Mean
Bart Kosko
and interquartile range (IQR) for dispersion. Other alternatives include trimming and Winsorising, as in the trimmed mean and the Winsorized mean.


See also

*
Estimation of covariance matrices In statistics, sometimes the covariance matrix of a multivariate random variable is not known but has to be estimated. Estimation of covariance matrices then deals with the question of how to approximate the actual covariance matrix on the basis o ...
* Scatter matrix * Unbiased estimation of standard deviation


References

{{Authority control Covariance and correlation Estimation methods Summary statistics Matrices U-statistics