HOME

TheInfoList



OR:

In multivariate statistics, if \varepsilon is a vector of n
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
s, and \Lambda is an n-dimensional symmetric matrix, then the
scalar Scalar may refer to: *Scalar (mathematics), an element of a field, which is used to define a vector space, usually the field of real numbers * Scalar (physics), a physical quantity that can be described by a single element of a number field such ...
quantity \varepsilon^T\Lambda\varepsilon is known as a quadratic form in \varepsilon.


Expectation

It can be shown that :\operatorname\left varepsilon^T\Lambda\varepsilon\right\operatorname\left Lambda \Sigma\right+ \mu^T\Lambda\mu where \mu and \Sigma are the
expected value In probability theory, the expected value (also called expectation, expectancy, mathematical expectation, mean, average, or first moment) is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a l ...
and
variance-covariance matrix In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements of ...
of \varepsilon, respectively, and tr denotes the trace of a matrix. This result only depends on the existence of \mu and \Sigma; in particular, normality of \varepsilon is ''not'' required. A book treatment of the topic of quadratic forms in random variables is that of Mathai and Provost.


Proof

Since the quadratic form is a scalar quantity, \varepsilon^T\Lambda\varepsilon = \operatorname(\varepsilon^T\Lambda\varepsilon). Next, by the cyclic property of the trace operator, : \operatorname operatorname(\varepsilon^T\Lambda\varepsilon)= \operatorname operatorname(\Lambda\varepsilon\varepsilon^T) Since the trace operator is a linear combination of the components of the matrix, it therefore follows from the linearity of the expectation operator that : \operatorname operatorname(\Lambda\varepsilon\varepsilon^T)= \operatorname(\Lambda \operatorname(\varepsilon\varepsilon^T)). A standard property of variances then tells us that this is : \operatorname(\Lambda (\Sigma + \mu \mu^T)). Applying the cyclic property of the trace operator again, we get : \operatorname(\Lambda\Sigma) + \operatorname(\Lambda \mu \mu^T) = \operatorname(\Lambda\Sigma) + \operatorname(\mu^T\Lambda\mu) = \operatorname(\Lambda\Sigma) + \mu^T\Lambda\mu.


Variance in the Gaussian case

In general, the variance of a quadratic form depends greatly on the distribution of \varepsilon. However, if \varepsilon ''does'' follow a multivariate normal distribution, the variance of the quadratic form becomes particularly tractable. Assume for the moment that \Lambda is a symmetric matrix. Then, :\operatorname \left varepsilon^T\Lambda\varepsilon\right= 2\operatorname\left Lambda \Sigma\Lambda \Sigma\right+ 4\mu^T\Lambda\Sigma\Lambda\mu. In fact, this can be generalized to find the covariance between two quadratic forms on the same \varepsilon (once again, \Lambda_1 and \Lambda_2 must both be symmetric): :\operatorname\left varepsilon^T\Lambda_1\varepsilon,\varepsilon^T\Lambda_2\varepsilon\right2\operatorname\left Lambda _1\Sigma\Lambda_2 \Sigma\right+ 4\mu^T\Lambda_1\Sigma\Lambda_2\mu. In addition, a quadratic form such as this follows a
generalized chi-squared distribution In probability theory and statistics, the generalized chi-squared distribution (or generalized chi-square distribution) is the distribution of a quadratic form of a multinormal variable (normal vector), or a linear combination of different no ...
.


Computing the variance in the non-symmetric case

Some texts incorrectly state that the above variance or covariance results hold without requiring \Lambda to be symmetric. The case for general \Lambda can be derived by noting that :\varepsilon^T\Lambda^T\varepsilon=\varepsilon^T\Lambda\varepsilon so :\varepsilon^T\tilde\varepsilon=\varepsilon^T\left(\Lambda+\Lambda^T\right)\varepsilon/2 is a quadratic form in the symmetric matrix \tilde=\left(\Lambda+\Lambda^T\right)/2, so the mean and variance expressions are the same, provided \Lambda is replaced by \tilde therein.


Examples of quadratic forms

In the setting where one has a set of observations y and an
operator matrix In statistics, the projection matrix (\mathbf), sometimes also called the influence matrix or hat matrix (\mathbf), maps the vector of response values (dependent variable values) to the vector of fitted values (or predicted values). It describes ...
H, then the
residual sum of squares In statistics, the residual sum of squares (RSS), also known as the sum of squared estimate of errors (SSE), is the sum of the squares of residuals (deviations predicted from actual empirical values of data). It is a measure of the discrepanc ...
can be written as a quadratic form in y: :\textrm=y^T(I-H)^T (I-H)y. For procedures where the matrix H is symmetric and idempotent, and the errors are Gaussian with covariance matrix \sigma^2I, \textrm/\sigma^2 has a chi-squared distribution with k degrees of freedom and noncentrality parameter \lambda, where :k=\operatorname\left I-H)^T(I-H)\right/math> :\lambda=\mu^T(I-H)^T(I-H)\mu/2 may be found by matching the first two central moments of a noncentral chi-squared random variable to the expressions given in the first two sections. If Hy estimates \mu with no bias, then the noncentrality \lambda is zero and \textrm/\sigma^2 follows a central chi-squared distribution.


See also

*
Quadratic form In mathematics, a quadratic form is a polynomial with terms all of degree two ("form" is another name for a homogeneous polynomial). For example, :4x^2 + 2xy - 3y^2 is a quadratic form in the variables and . The coefficients usually belong to a ...
*
Covariance matrix In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements of ...
* Matrix representation of conic sections


References

{{DEFAULTSORT:Quadratic Form (Statistics) Statistical theory Quadratic forms