Principle Of Marginality
   HOME

TheInfoList



OR:

In
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, the principle of marginality, sometimes called hierarchical principle, is the fact that the average (or main) effects of variables in an analysis are marginal to their interaction effect—that is, the main effect of one
explanatory variable A variable is considered dependent if it depends on (or is hypothesized to depend on) an independent variable. Dependent variables are studied under the supposition or demand that they depend, by some law or rule (e.g., by a mathematical function ...
captures the effect of that variable averaged over all values of a second explanatory variable whose value influences the first variable's effect. The principle of marginality implies that, in general, it is wrong to test, estimate, or interpret main effects of explanatory variables where the variables interact or, similarly, to model interaction effects but delete main effects that are marginal to them.Fox, J
Regression Notes
While such models are interpretable, they lack applicability, as they ignore the dependence of a variable's effect upon another variable's value. Nelder and VenablesVenables, W.N. (1998)
"Exegeses on Linear Models"
Paper presented to the S-PLUS User's Conference Washington, DC, 8–9 October 1998.
have argued strongly for the importance of this principle in regression analysis.


Regression form

If two independent continuous variables, say ''x'' and ''z'', both influence a
dependent variable A variable is considered dependent if it depends on (or is hypothesized to depend on) an independent variable. Dependent variables are studied under the supposition or demand that they depend, by some law or rule (e.g., by a mathematical functio ...
''y'', and if the extent of the effect of each independent variable depends on the level of the other independent variable then the regression equation can be written as: :y_i=a+bx_i+cz_i+d(x_iz_i)+ e_i, where ''i'' indexes observations, ''a'' is the intercept term, ''b'', ''c'', and ''d'' are
effect size In statistics, an effect size is a value measuring the strength of the relationship between two variables in a population, or a sample-based estimate of that quantity. It can refer to the value of a statistic calculated from a sample of data, the ...
parameters to be estimated, and ''e'' is the
error An error (from the Latin , meaning 'to wander'Oxford English Dictionary, s.v. “error (n.), Etymology,” September 2023, .) is an inaccurate or incorrect action, thought, or judgement. In statistics, "error" refers to the difference between t ...
term. If this is the correct model, then the omission of any of the right-side terms would be incorrect, resulting in misleading interpretation of the regression results. With this model, the effect of ''x'' upon ''y'' is given by the
partial derivative In mathematics, a partial derivative of a function of several variables is its derivative with respect to one of those variables, with the others held constant (as opposed to the total derivative, in which all variables are allowed to vary). P ...
of ''y'' with respect to ''x''; this is b+dz_i, which depends on the specific value z_i at which the partial derivative is being evaluated. Hence, the main effect of ''x'' – the effect averaged over all values of ''z'' – is meaningless as it depends on the design of the experiment (specifically on the relative frequencies of the various values of ''z'') and not just on the underlying relationships. Hence: * In the case of interaction, it is wrong to try to test, estimate, or interpret a "main effect" coefficient ''b'' or ''c'', omitting the interaction term. In addition: * In the case of interaction, it is wrong to not include ''b'' or ''c'', because this will give incorrect estimates of the interaction.The above regression model, with two independent continuous variables, is presented with a numerical example, in Stata, as Case 3 i
What happens if you omit the main effect in a regression model with an interaction?


See also

*
General linear model The general linear model or general multivariate regression model is a compact way of simultaneously writing several multiple linear regression models. In that sense it is not a separate statistical linear model. The various multiple linear regre ...
*
Analysis of variance Analysis of variance (ANOVA) is a family of statistical methods used to compare the Mean, means of two or more groups by analyzing variance. Specifically, ANOVA compares the amount of variation ''between'' the group means to the amount of variati ...


References

{{Reflist Marginality Regression analysis Analysis of variance