Factorial Experiments
   HOME

TheInfoList



OR:

In
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, a factorial experiment (also known as full factorial experiment) investigates how multiple factors influence a specific outcome, called the
response variable A variable is considered dependent if it depends on (or is hypothesized to depend on) an independent variable. Dependent variables are studied under the supposition or demand that they depend, by some law or rule (e.g., by a mathematical function ...
. Each factor is tested at distinct values, or levels, and the
experiment An experiment is a procedure carried out to support or refute a hypothesis, or determine the efficacy or likelihood of something previously untried. Experiments provide insight into cause-and-effect by demonstrating what outcome occurs whe ...
includes every possible combination of these levels across all factors. This comprehensive approach lets researchers see not only how each factor individually affects the response, but also how the factors
interact Advocates for Informed Choice, dba interACT or interACT Advocates for Intersex Youth, is a 501(c)(3) nonprofit organization advocating for the legal and human rights of children with intersex traits. The organization was founded in 2006 and fo ...
and influence each other. Often, factorial experiments simplify things by using just two levels for each factor. A 2x2 factorial design, for instance, has two factors, each with two levels, leading to four unique combinations to test. The interaction between these factors is often the most crucial finding, even when the individual factors also have an effect. If a full factorial design becomes too complex due to the sheer number of combinations, researchers can use a
fractional factorial design In statistics, a fractional factorial design is a way to conduct experiments with fewer experimental runs than a full factorial design. Instead of testing every single combination of factors, it tests only a carefully selected portion. This " ...
. This method strategically omits some combinations (usually at least half) to make the experiment more manageable. These combinations of factor levels are sometimes called ''runs'' (of an experiment), ''points'' (viewing the combinations as vertices of a graph), and ''cells'' (arising as intersections of rows and columns).


History

Factorial designs were used in the 19th century by
John Bennet Lawes Sir John Bennet Lawes, 1st Baronet, FRS (28 December 1814 – 31 August 1900) was an English entrepreneur and agricultural scientist. He founded an experimental farm at his home at Rothamsted Manor that eventually became Rothamsted Research, ...
and
Joseph Henry Gilbert Sir Joseph Henry Gilbert (1 August 1817 – 23 December 1901) was an English chemist, noteworthy for his long career spent improving the methods of practical agriculture. Along with J.B. Lawes, he conducted experiments at Rothamstead for forty ...
of the
Rothamsted Experimental Station Rothamsted Research, previously known as the Rothamsted Experimental Station and then the Institute of Arable Crops Research, is one of the oldest agricultural research institutions in the world, having been founded in 1843. It is located at Harp ...
.
Ronald Fisher Sir Ronald Aylmer Fisher (17 February 1890 – 29 July 1962) was a British polymath who was active as a mathematician, statistician, biologist, geneticist, and academic. For his work in statistics, he has been described as "a genius who a ...
argued in 1926 that "complex" designs (such as factorial designs) were more efficient than studying one factor at a time. Fisher wrote, A factorial design allows the effect of several factors and even interactions between them to be determined with the same number of trials as are necessary to determine any one of the effects by itself with the same degree of accuracy.
Frank Yates Frank Yates FRS (12 May 1902 – 17 June 1994) was one of the pioneers of 20th-century statistics. Biography Yates was born in Manchester, England, the eldest of five children (and only son) of seed merchant and botanist Percy Yates and ...
made significant contributions, particularly in the analysis of designs, by the Yates analysis. The term "factorial" may not have been used in print before 1935, when Fisher used it in his book ''
The Design of Experiments ''The Design of Experiments'' is a 1935 book by the English statistician Ronald Fisher about the design of experiments and is considered a foundational work in experimental design. Among other contributions, the book introduced the concept of th ...
''.


Advantages and disadvantages of factorial experiments

Many people examine the effect of only a single factor or variable. Compared to such one-factor-at-a-time (OFAT) experiments, factorial experiments offer several advantages * Factorial designs are more efficient than OFAT experiments. They provide more information at similar or lower cost. They can find optimal conditions faster than OFAT experiments. * When the effect of one factor is different for different levels of another factor, it cannot be detected by an OFAT experiment design. Factorial designs are required to detect such interactions. Use of OFAT when interactions are present can lead to serious misunderstanding of how the response changes with the factors. * Factorial designs allow the effects of a factor to be estimated at several levels of the other factors, yielding conclusions that are valid over a range of experimental conditions. The main disadvantage of the full factorial design is its sample size requirement, which grows exponentially with the number of factors or inputs considered. Alternative strategies with improved computational efficiency include
fractional factorial design In statistics, a fractional factorial design is a way to conduct experiments with fewer experimental runs than a full factorial design. Instead of testing every single combination of factors, it tests only a carefully selected portion. This " ...
s,
Latin hypercube sampling Latin hypercube sampling (LHS) is a statistical method for generating a near-random sample of parameter values from a multidimensional distribution. The sampling method is often used to construct computer experiments or for Monte Carlo integratio ...
, and quasi-random sampling techniques.


Example of advantages of factorial experiments

In his book, ''Improving Almost Anything: Ideas and Essays'', statistician
George Box George Edward Pelham Box (18 October 1919 – 28 March 2013) was a British statistician, who worked in the areas of quality control, time-series analysis, design of experiments, and Bayesian inference. He has been called "one of the gre ...
gives many examples of the benefits of factorial experiments. Here is one. Engineers at the bearing manufacturer SKF wanted to know if changing to a less expensive "cage" design would affect bearing lifespan. The engineers asked Christer Hellstrand, a statistician, for help in designing the experiment. Box reports the following. "The results were assessed by an accelerated life test. … The runs were expensive because they needed to be made on an actual production line and the experimenters were planning to make four runs with the standard cage and four with the modified cage. Christer asked if there were other factors they would like to test. They said there were, but that making added runs would exceed their budget. Christer showed them how they could test two additional factors "for free" – without increasing the number of runs and without reducing the accuracy of their estimate of the cage effect. In this arrangement, called a 2×2×2 factorial design, each of the three factors would be run at two levels and all the eight possible combinations included. The various combinations can conveniently be shown as the vertices of a cube ... " "In each case, the standard condition is indicated by a minus sign and the modified condition by a plus sign. The factors changed were heat treatment, outer ring osculation, and cage design. The numbers show the relative lengths of lives of the bearings. If you look at he cube plot you can see that the choice of cage design did not make a lot of difference. … But, if you average the pairs of numbers for cage design, you get the able below which shows what the two other factors did. … It led to the extraordinary discovery that, in this particular application, the life of a bearing can be increased fivefold if the two factor(s) outer ring osculation and inner ring heat treatments are increased together." "Remembering that bearings like this one have been made for decades, it is at first surprising that it could take so long to discover so important an improvement. A likely explanation is that, because most engineers have, until recently, employed only one factor at a time experimentation, interaction effects have been missed."


Example

The simplest factorial experiment contains two levels for each of two factors. Suppose an engineer wishes to study the total power used by each of two different motors, A and B, running at each of two different speeds, 2000 or 3000 RPM. The factorial experiment would consist of four experimental units: motor A at 2000 RPM, motor B at 2000 RPM, motor A at 3000 RPM, and motor B at 3000 RPM. Each combination of a single level selected from every factor is present once. This experiment is an example of a 22 (or 2×2) factorial experiment, so named because it considers two levels (the base) for each of two factors (the power or superscript), or #levels#factors, producing 22=4 factorial points. Designs can involve many independent variables. As a further example, the effects of three input variables can be evaluated in eight experimental conditions shown as the corners of a cube. This can be conducted with or without replication, depending on its intended purpose and available resources. It will provide the effects of the three independent variables on the dependent variable and possible interactions.


Notation

Factorial experiments are described by two things: the number of factors, and the number of levels of each factor. For example, a 2×3 factorial experiment has two factors, the first at 2 levels and the second at 3 levels. Such an experiment has 2×3=6 treatment combinations or cells. Similarly, a 2×2×3 experiment has three factors, two at 2 levels and one at 3, for a total of 12 treatment combinations. If every factor has ''s'' levels (a so-called ''fixed-level'' or ''symmetric'' design), the experiment is typically denoted by ''sk'', where ''k'' is the number of factors. Thus a 25 experiment has 5 factors, each at 2 levels. Experiments that are not fixed-level are said to be ''mixed-level'' or ''asymmetric''. There are various traditions to denote the levels of each factor. If a factor already has natural units, then those are used. For example, a shrimp aquaculture experiment might have factors ''temperature'' at 25 °C and 35 °C, ''density'' at 80 or 160 shrimp/40 liters, and ''salinity'' at 10%, 25% and 40%. In many cases, though, the factor levels are simply categories, and the coding of levels is somewhat arbitrary. For example, the levels of an 6-level factor might simply be denoted 1, 2, ..., 6. Treatment combinations are denoted by ordered pairs or, more generally, ordered
tuple In mathematics, a tuple is a finite sequence or ''ordered list'' of numbers or, more generally, mathematical objects, which are called the ''elements'' of the tuple. An -tuple is a tuple of elements, where is a non-negative integer. There is o ...
s. In the aquaculture experiment, the ordered triple (25, 80, 10) represents the treatment combination having the lowest level of each factor. In a general 2×3 experiment the ordered pair (2, 1) would indicate the cell in which factor ''A'' is at level 2 and factor ''B'' at level 1. The parentheses are often dropped, as shown in the accompanying table. To denote factor levels in 2''k'' experiments, three particular systems appear in the literature: * The values 1 and 0; * the values 1 and −1, often simply abbreviated by + and −; * A lower-case letter with the exponent 0 or 1. If these values represent "low" and "high" settings of a treatment, then it is natural to have 1 represent "high", whether using 0 and 1 or −1 and 1. This is illustrated in the accompanying table for a 2×2 experiment. If the factor levels are simply categories, the correspondence might be different; for example, it is natural to represent "control" and "experimental" conditions by coding "control" as 0 if using 0 and 1, and as 1 if using 1 and −1. An example of the latter is given
below Below may refer to: *Earth *Ground (disambiguation) *Soil *Floor * Bottom (disambiguation) *Less than *Temperatures below freezing *Hell or underworld People with the surname * Ernst von Below (1863–1955), German World War I general * Fred Belo ...
. That example illustrates another use of the coding +1 and −1. For other fixed-level (''sk'') experiments, the values 0, 1, ..., ''s''−1 are often used to denote factor levels. These are the values of the integers
modulo In computing and mathematics, the modulo operation returns the remainder or signed remainder of a division, after one number is divided by another, the latter being called the '' modulus'' of the operation. Given two positive numbers and , mo ...
''s'' when ''s'' is prime.This choice of factor levels facilitates the use of algebra to handle certain issues of experimental design. If ''s'' is a power of a prime, the levels may be denoted by the elements of the
finite field In mathematics, a finite field or Galois field (so-named in honor of Évariste Galois) is a field (mathematics), field that contains a finite number of Element (mathematics), elements. As with any field, a finite field is a Set (mathematics), s ...
''GF(s)'' for the same reason.


Contrasts, main effects and interactions

The ''expected response'' to a given treatment combination is called a ''cell mean'', usually denoted using the Greek letter μ. (The term ''cell'' is borrowed from its use in tables of data.) This notation is illustrated here for the 2 × 3 experiment. A ''
contrast Contrast may refer to: Science * Contrast (vision), the contradiction in form, colour and light between parts of an image * Contrast (statistics), a combination of averages whose coefficients add up to zero, or the difference between two means * ...
in cell means'' is a
linear combination In mathematics, a linear combination or superposition is an Expression (mathematics), expression constructed from a Set (mathematics), set of terms by multiplying each term by a constant and adding the results (e.g. a linear combination of ''x'' a ...
of cell means in which the coefficients sum to 0. Contrasts are of interest in themselves, and are the building blocks by which main effects and interactions are defined. In the 2 × 3 experiment illustrated here, the expression
\mu_ - \mu_
is a contrast that compares the mean responses of the treatment combinations 11 and 12. (The coefficients here are 1 and –1.) The contrast
\mu_ + \mu_ + \mu_ - \mu_ - \mu_ - \mu_
is said to ''belong to the main effect of factor A'' as it contrasts the responses to the "1" level of factor A with those for the "2" level. The main effect of ''A'' is said to be ''absent'' if the true values of the cell means \mu_ make this expression equal to 0. Since the true cell means are unobservable in principle, a
statistical hypothesis test A statistical hypothesis test is a method of statistical inference used to decide whether the data provide sufficient evidence to reject a particular hypothesis. A statistical hypothesis test typically involves a calculation of a test statistic. T ...
is used to assess whether this expression equals 0. ''Interaction'' in a factorial experiment is the lack of additivity between factors, and is also expressed by contrasts. In the 2 × 3 experiment, the contrasts
\mu_ - \mu_ - \mu_ + \mu_   and   \mu_ - \mu_ - \mu_ + \mu_
''belong to the A × B interaction''; interaction is ''absent'' (additivity is ''present'') if these expressions equal 0. Additivity may be viewed as a kind of parallelism between factors, as illustrated in the Analysis section
below Below may refer to: *Earth *Ground (disambiguation) *Soil *Floor * Bottom (disambiguation) *Less than *Temperatures below freezing *Hell or underworld People with the surname * Ernst von Below (1863–1955), German World War I general * Fred Belo ...
. As with main effects, one assesses the assumption of additivity by performing a hypothesis test. Since it is the coefficients of these contrasts that carry the essential information, they are often displayed as
column vector In linear algebra, a column vector with elements is an m \times 1 matrix consisting of a single column of entries, for example, \boldsymbol = \begin x_1 \\ x_2 \\ \vdots \\ x_m \end. Similarly, a row vector is a 1 \times n matrix for some , c ...
s. For the example above, such a table might look like this: The columns of such a table are called ''contrast vectors'': their components add up to 0. Each effect is determined by both the ''pattern of components'' in its columns and the ''number of columns''. The patterns of components of these columns reflect the general definitions given by
Bose Bose may refer to: * Bose (crater), a lunar crater. * ''Bose'' (film), a 2004 Indian Tamil film starring Srikanth and Sneha * Bose (surname), a surname (and list of people with the name) * Bose (given name), a given name * Bose, Italy, a ''fra ...
: * A contrast vector ''belongs to the main effect of a particular factor'' if the values of its components depend only on the level of that factor. * A contrast vector ''belongs to the interaction of two factors'', say ''A'' and ''B'', if (i) the values of its components depend only on the levels of ''A'' and ''B'', and (ii) it is
orthogonal In mathematics, orthogonality (mathematics), orthogonality is the generalization of the geometric notion of ''perpendicularity''. Although many authors use the two terms ''perpendicular'' and ''orthogonal'' interchangeably, the term ''perpendic ...
(perpendicular) to the contrast vectors representing the main effects of ''A'' and ''B''. Orthogonality is determined by computing the
dot product In mathematics, the dot product or scalar productThe term ''scalar product'' means literally "product with a Scalar (mathematics), scalar as a result". It is also used for other symmetric bilinear forms, for example in a pseudo-Euclidean space. N ...
of vectors.
Similar definitions hold for interactions of more than two factors. In the 2 × 3 example, for instance, the pattern of the ''A'' column follows the pattern of the levels of factor ''A'', indicated by the first component of each cell. Similarly, the pattern of the ''B'' columns follows the levels of factor ''B'' (sorting on ''B'' makes this easier to see). The number of columns needed to specify each effect is the ''degrees of freedom'' for the effect, and is an essential quantity in the
analysis of variance Analysis of variance (ANOVA) is a family of statistical methods used to compare the Mean, means of two or more groups by analyzing variance. Specifically, ANOVA compares the amount of variation ''between'' the group means to the amount of variati ...
. The formula is as follows: * A main effect for a factor with ''s'' levels has ''s''−1 degrees of freedom. * The interaction of two factors with ''s''1 and ''s''2 levels, respectively, has (''s''1−1)(''s''2−1) degrees of freedom. The formula for more than two factors follows this pattern. In the 2 × 3 example above, the degrees of freedom for the two main effects and the interaction — the number of columns for each — are 1, 2 and 2, respectively.


Examples

In the tables in the following examples, the entries in the "cell" column are treatment combinations: The first component of each combination is the level of factor ''A'', the second for factor ''B'', and the third (in the 2 × 2 × 2 example) the level of factor ''C''. The entries in each of the other columns sum to 0, so that each column is a contrast vector. A 3 × 3 experiment: Here we expect 3-1 = 2 degrees of freedom each for the main effects of factors ''A'' and ''B'', and (3-1)(3-1) = 4 degrees of freedom for the ''A × B'' interaction. This accounts for the number of columns for each effect in the accompanying table. The two contrast vectors for ''A'' depend only on the level of factor ''A''. This can be seen by noting that the pattern of entries in each ''A'' column is the same as the pattern of the first component of "cell". (If necessary, sorting the table on ''A'' will show this.) Thus these two vectors belong to the main effect of ''A''. Similarly, the two contrast vectors for ''B'' depend only on the level of factor ''B'', namely the second component of "cell", so they belong to the main effect of ''B''. The last four column vectors belong to the ''A × B'' interaction, as their entries depend on the values of both factors, and as all four columns are orthogonal to the columns for ''A'' and ''B''. The latter can be verified by taking
dot product In mathematics, the dot product or scalar productThe term ''scalar product'' means literally "product with a Scalar (mathematics), scalar as a result". It is also used for other symmetric bilinear forms, for example in a pseudo-Euclidean space. N ...
s. A 2 × 2 × 2 experiment: This will have 1 degree of freedom for every main effect and interaction. For example, a two-factor interaction will have (2-1)(2-1) = 1 degree of freedom. Thus just a single column is needed to specify each of the seven effects. The columns for ''A'', ''B'' and ''C'' represent the corresponding main effects, as the entries in each column depend only on the level of the corresponding factor. For example, the entries in the ''B'' column follow the same pattern as the middle component of "cell", as can be seen by sorting on ''B''. The columns for ''AB'', ''AC'' and ''BC'' represent the corresponding two-factor interactions. For example, (i) the entries in the ''BC'' column depend on the second and third (''B'' and ''C'') components of ''cell'', and are independent of the first (''A'') component, as can be seen by sorting on ''BC''; and (ii) the ''BC'' column is orthogonal to columns ''B'' and ''C'', as can be verified by computing dot products. Finally, the ''ABC'' column represents the three-factor interaction: its entries depend on the levels of all three factors, and it is orthogonal to the other six contrast vectors. Combined and read row-by-row, columns ''A, B, C'' give an alternate notation, mentioned above, for the treatment combinations (cells) in this experiment: cell 000 corresponds to +++, 001 to ++−, etc. In columns ''A'' through ''ABC'', the number 1 may be replaced by any constant, because the resulting columns will still be contrast vectors. For example, it is common to use the number 1/4 in 2 × 2 × 2 experiments And 1/2k-1 in 2k experiments. to define each main effect or interaction, and to declare, for example, that the contrast
(\mu_ + \mu_ + \mu_ + \mu_)/4 - (\mu_ + \mu_ + \mu_ + \mu_)/4
is "the" main effect of factor ''A'', a numerical quantity that can be estimated.


Implementation

For more than two factors, a 2''k'' factorial experiment can usually be recursively designed from a 2''k''−1 factorial experiment by replicating the 2''k''−1 experiment, assigning the first replicate to the first (or low) level of the new factor, and the second replicate to the second (or high) level. This framework can be generalized to, ''e.g.'', designing three replicates for three level factors, ''etc''. A factorial experiment allows for estimation of
experimental error Observational error (or measurement error) is the difference between a measured value of a quantity and its unknown true value.Dodge, Y. (2003) ''The Oxford Dictionary of Statistical Terms'', OUP. Such errors are inherent in the measurement pr ...
in two ways. The experiment can be replicated, or the sparsity-of-effects principle can often be exploited. Replication is more common for small experiments and is a very reliable way of assessing experimental error. When the number of factors is large (typically more than about 5 factors, but this does vary by application), replication of the design can become operationally difficult. In these cases, it is common to only run a single replicate of the design, and to assume that factor interactions of more than a certain order (say, between three or more factors) are negligible. Under this assumption, estimates of such high order interactions are estimates of an exact zero, thus really an estimate of experimental error. When there are many factors, many experimental runs will be necessary, even without replication. For example, experimenting with 10 factors at two levels each produces 210=1024 combinations. At some point this becomes infeasible due to high cost or insufficient resources. In this case,
fractional factorial designs In statistics, a fractional factorial design is a way to Design of experiments, conduct experiments with fewer experimental runs than a full factorial design. Instead of testing every single combination of factors, it tests only a carefully sel ...
may be used. As with any statistical experiment, the experimental runs in a factorial experiment should be randomized to reduce the impact that
bias Bias is a disproportionate weight ''in favor of'' or ''against'' an idea or thing, usually in a way that is inaccurate, closed-minded, prejudicial, or unfair. Biases can be innate or learned. People may develop biases for or against an individ ...
could have on the experimental results. In practice, this can be a large operational challenge. Factorial experiments can be used when there are more than two levels of each factor. However, the number of experimental runs required for three-level (or more) factorial designs will be considerably greater than for their two-level counterparts. Factorial designs are therefore less attractive if a researcher wishes to consider more than two levels.


Analysis

A factorial experiment can be analyzed using
ANOVA Analysis of variance (ANOVA) is a family of statistical methods used to compare the means of two or more groups by analyzing variance. Specifically, ANOVA compares the amount of variation ''between'' the group means to the amount of variation ''w ...
or regression analysis. To compute the main effect of a factor "A" in a 2-level experiment, subtract the average response of all experimental runs for which A was at its low (or first) level from the average response of all experimental runs for which A was at its high (or second) level. Other useful exploratory analysis tools for factorial experiments include
main effect In the design of experiments and analysis of variance, a main effect is the effect of an independent variable on a dependent variable averaged across the levels of any other independent variables. The term is frequently used in the context of fact ...
s plots, interaction plots, Pareto plots, and a
normal probability plot The normal probability plot is a graphical technique to identify substantive departures from normality. This includes identifying outliers, skewness, kurtosis, a need for transformations, and mixtures. Normal probability plots are made of raw ...
of the estimated effects. When the factors are continuous, two-level factorial designs assume that the effects are
linear In mathematics, the term ''linear'' is used in two distinct senses for two different properties: * linearity of a '' function'' (or '' mapping''); * linearity of a '' polynomial''. An example of a linear function is the function defined by f(x) ...
. If a
quadratic In mathematics, the term quadratic describes something that pertains to squares, to the operation of squaring, to terms of the second degree, or equations or formulas that involve such terms. ''Quadratus'' is Latin for ''square''. Mathematics ...
effect is expected for a factor, a more complicated experiment should be used, such as a central composite design. Optimization of factors that could have quadratic effects is the primary goal of
response surface methodology In statistics, response surface methodology (RSM) explores the relationships between several explanatory variables and one or more response variables. RSM is an empirical model which employs the use of mathematical and statistical techniques to r ...
.


Analysis example

Montgomery gives the following example of analysis of a factorial experiment:.
An engineer would like to increase the filtration rate (output) of a process to produce a chemical, and to reduce the amount of
formaldehyde Formaldehyde ( , ) (systematic name methanal) is an organic compound with the chemical formula and structure , more precisely . The compound is a pungent, colourless gas that polymerises spontaneously into paraformaldehyde. It is stored as ...
used in the process. Previous attempts to reduce the formaldehyde have lowered the filtration rate. The current filtration rate is 75 gallons per hour. Four factors are considered: temperature (A), pressure (B), formaldehyde concentration (C), and stirring rate (D). Each of the four factors will be tested at two levels.
Onwards, the minus (−) and plus (+) signs will indicate whether the factor is run at a low or high level, respectively. File:Montgomery filtration rates.svg, Plot of the main effects showing the filtration rates for the low (−) and high (+) settings for each factor. File:Interaction plots filtration rate.png, Plot of the interaction effects showing the mean filtration rate at each of the four possible combinations of levels for a given pair of factors. The non-parallel lines in the A:C interaction plot indicate that the effect of factor A depends on the level of factor C. A similar results holds for the A:D interaction. The graphs indicate that factor B has little effect on filtration rate. The
analysis of variance Analysis of variance (ANOVA) is a family of statistical methods used to compare the Mean, means of two or more groups by analyzing variance. Specifically, ANOVA compares the amount of variation ''between'' the group means to the amount of variati ...
(ANOVA) including all 4 factors and all possible interaction terms between them yields the coefficient estimates shown in the table below. Because there are 16 observations and 16 coefficients (intercept, main effects, and interactions),
p-value In null-hypothesis significance testing, the ''p''-value is the probability of obtaining test results at least as extreme as the result actually observed, under the assumption that the null hypothesis is correct. A very small ''p''-value means ...
s cannot be calculated for this model. The coefficient values and the graphs suggest that the important factors are A, C, and D, and the interaction terms A:C and A:D. The coefficients for A, C, and D are all positive in the ANOVA, which would suggest running the process with all three variables set to the high value. However, the main effect of each variable is the average over the levels of the other variables. The A:C interaction plot above shows that the effect of factor A depends on the level of factor C, and vice versa. Factor A (temperature) has very little effect on filtration rate when factor C is at the + level. But Factor A has a large effect on filtration rate when factor C (formaldehyde) is at the − level. The combination of A at the + level and C at the − level gives the highest filtration rate. This observation indicates how one-factor-at-a-time analyses can miss important interactions. Only by varying both factors A and C at the same time could the engineer discover that the effect of factor A depends on the level of factor C. The best filtration rate is seen when A and D are at the high level, and C is at the low level. This result also satisfies the objective of reducing formaldehyde (factor C). Because B does not appear to be important, it can be dropped from the model. Performing the ANOVA using factors A, C, and D, and the interaction terms A:C and A:D, gives the result shown in the following table, in which all the terms are significant (p-value < 0.05).


See also

*
Combinatorial design Combinatorial design theory is the part of combinatorial mathematics that deals with the existence, construction and properties of systems of finite sets whose arrangements satisfy generalized concepts of ''balance'' and/or ''symmetry''. These co ...
*
Design of experiments The design of experiments (DOE), also known as experiment design or experimental design, is the design of any task that aims to describe and explain the variation of information under conditions that are hypothesized to reflect the variation. ...
*
Orthogonal array In mathematics, an orthogonal array (more specifically, a fixed-level orthogonal array) is a "table" (array) whose entries come from a fixed finite set of symbols (for example, ), arranged in such a way that there is an integer ''t'' so that for ev ...
* Plackett–Burman design *
Taguchi methods Taguchi methods () are statistical methods, sometimes called robust design methods, developed by Genichi Taguchi to improve the quality of manufactured goods, and more recently also applied to engineering, biotechnology, marketing and advertising. ...
*
Welch's t-test In statistics, Welch's ''t''-test, or unequal variances ''t''-test, is a two-sample location test which is used to test the (null) hypothesis that two populations have equal means. It is named for its creator, Bernard Lewis Welch, and is an adapt ...


Explanatory footnotes


Notes


References

* * * * * * * * * * *


External links


Factorial Designs (California State University, Fresno)GOV.UK Factorial randomised controlled trials (Public Health England)
{{Statistics, collection Design of experiments Statistical process control