White Test
   HOME

TheInfoList



OR:

White test is a
statistical test A statistical hypothesis test is a method of statistical inference used to decide whether the data provide sufficient evidence to reject a particular hypothesis. A statistical hypothesis test typically involves a calculation of a test statistic. ...
that establishes whether the
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
of the errors in a regression model is constant: that is for homoskedasticity. This test, and an estimator for
heteroscedasticity-consistent standard errors The topic of heteroskedasticity-consistent (HC) standard errors arises in statistics and econometrics in the context of linear regression and time series analysis. These are also known as heteroskedasticity-robust standard errors (or simply robust ...
, were proposed by Halbert White in 1980. These methods have become widely used, making this paper one of the most cited articles in economics. In cases where the White test statistic is statistically significant, heteroskedasticity may not necessarily be the cause; instead the problem could be a specification error. In other words, the White test can be a test of heteroskedasticity or specification error or both. If no cross product terms are introduced in the White test procedure, then this is a test of pure heteroskedasticity. If cross products are introduced in the model, then it is a test of both heteroskedasticity and specification bias.


Testing constant variance

To test for constant variance one undertakes an auxiliary regression analysis: this regresses the squared residuals from the original regression model onto a set of
regressor A variable is considered dependent if it depends on (or is hypothesized to depend on) an independent variable. Dependent variables are studied under the supposition or demand that they depend, by some law or rule (e.g., by a mathematical function ...
s that contain the original regressors along with their squares and cross-products. One then inspects the ''R''2. The Lagrange multiplier (LM) test statistic is the product of the ''R''2 value and sample size: :\text = n R^2 . This follows a
chi-squared distribution In probability theory and statistics, the \chi^2-distribution with k Degrees of freedom (statistics), degrees of freedom is the distribution of a sum of the squares of k Independence (probability theory), independent standard normal random vari ...
, with degrees of freedom equal to ''P'' − 1, where ''P'' is the number of estimated parameters (in the auxiliary regression). The logic of the test is as follows. First, the squared residuals from the original model serve as a proxy for the variance of the error term at each observation. (The error term is assumed to have a mean of zero, and the
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
of a zero-mean random variable is just the expectation of its square.) The independent variables in the auxiliary regression account for the possibility that the error variance depends on the values of the original regressors in some way (linear or quadratic). If the error term in the original model is in fact homoskedastic (has a constant variance) then the coefficients in the auxiliary regression (besides the constant) should be statistically indistinguishable from zero and the ''R''2 should be “small". Conversely, a “large" ''R''2 (scaled by the sample size so that it follows the chi-squared distribution) counts against the hypothesis of homoskedasticity. An alternative to the White test is the Breusch–Pagan test, where the Breusch-Pagan test is designed to detect only linear forms of heteroskedasticity. Under certain conditions and a modification of one of the tests, they can be found to be algebraically equivalent. If homoskedasticity is rejected one can use
heteroskedasticity-consistent standard errors The topic of heteroskedasticity-consistent (HC) standard errors arises in statistics and econometrics in the context of linear regression and time series analysis. These are also known as heteroskedasticity-robust standard errors (or simply robust ...
.


Software implementations

* In R, White's Test can be implemented using the white function of the skedastic package. * In
Python Python may refer to: Snakes * Pythonidae, a family of nonvenomous snakes found in Africa, Asia, and Australia ** ''Python'' (genus), a genus of Pythonidae found in Africa and Asia * Python (mythology), a mythical serpent Computing * Python (prog ...
, White's Test can be implemented using the het_white function of the statsmodels.stats.diagnostic.het_white *In
Stata Stata (, , alternatively , occasionally stylized as STATA) is a general-purpose Statistics, statistical software package developed by StataCorp for data manipulation, visualization, statistics, and automated reporting. It is used by researchers ...
, the test can be implemented using the estat imtest, white function.


See also

*
Heteroskedasticity In statistics, a sequence of random variables is homoscedastic () if all its random variables have the same finite variance; this is also known as homogeneity of variance. The complementary notion is called heteroscedasticity, also known as hete ...
* Breusch–Pagan test * Glejser test * Goldfeld–Quandt test * Park test


References


Further reading

* * * {{cite book , last=Wooldridge , first=Jeffrey M. , author-link=Jeffrey Wooldridge , title=Introductory Econometrics: A Modern Approach , publisher=South-Western , edition=Fifth , year=2013 , isbn=978-1-111-53439-4 , pages=269–70 Statistical tests Regression diagnostics