Chapter 11

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 18

Chapter 11: Factorial Designs

11.1 Two factor factorial designs 22 (2 levels 2 factors )


This situation is similar to the randomized block design from the previous
chapter. However, in addition to the effects within the groups (A) and within
the blocks (B), we now also consider so-called interaction effects between the
two parameters A & B.
Schematically speaking we have:

with n = # data points; c = # of groups (factor A); r = # of groups (factor


B); and n` = # of repeat experiments.
We now calculate the various sum of squares:
r c n'

∑∑∑ X
( )
r c n' i jk

SST = ∑∑∑ X i jk − X
2
i =1 j =1 k =1
sum of squares total; with X =
i =1 j =1 k =1 r c n'
c n'

∑∑ X
( )
r i jk

sum of squares factor A; SSA = c n' ∑ X i •• − X


2
j =1 k =1
with X i •• =
i =1 c n'
r n'

∑∑ X
( )
c i jk
sum of squares factor B; SSB = r n' ∑ X • j • − X
2
with X • j • = i =1 k =1

j =1 r n'

sum of squares interaction AB;


r c
(
SSAB = n' ∑∑ X i j• − X i•• − X • j • + X
i =1 j =1
)
2

Random error
( ) X i jk
r c n' n'
SSE = ∑∑∑ X i jk − X ij• with X ij• = ∑
2
sum of squares error;
i =1 j =1 k =1 k =1 n'

The mean sum of squares is derived by dividing the values by the


corresponding degree of freedom.
The present case is called the "Two-Factor ANOVA with Replications"

You may realize that the present calculations are very similar to the
randomized block situation in the previous chapter.
However, please realize that it is necessary for this type of analysis to have
repeat measurements. Otherwise MSE will become infinite and no F-ratios
can be calculated.
The textbook by Montgomery had another way of simplifying the actual
calculation of the sum of squares.
You may want to decide for yourself if one formula is easier than the other.
Use Laundry2 excel file as an example.
(manipulate the data to show weak and strong AB-interactions)
The AB-interaction effects are sometimes difficult to see in a table format.
Thus, graphical representations might be more appropriate.
Plot the averages for the various repeat groups. Parallel lines indicate a lack of
interaction, whereas a strongly different slope or even crossing lines are
indications of strong interactions.

Similar to the previous chapter we can run the Tukey-Kramer or the Fischer-
LSD test to evaluate individual differences between multiple levels within one
factor (more than 2 levels). !! remember to use the correct degree of freedoms!!
Depending on how the levels of one or both factors have been selected, we
have to use different processes for our F-ratio analysis! (see following table)
(Excel assumes fixed-levels)
Residual Analysis:
We test the experimental data against the assumed model (fit-equation).
In case the model describes the situation correctly and there are no 'outlier',
the residuals should follow a straight line in a normal probability plot.
We will use the data from "Laundry2" as an example.
1st, we have to define our model (fit-equation)
The ANOVA analysis indicated that the AB-interaction effect is relative
minor. Thus, we will first consider a simple linear effect of A, detergent brand,
and B, temperature.
Warm Hot predicted values residuals Warm Hot

X 14 18 XA + XB − X X -1.3 -0.1
X 16 19 Average: X 0.7 0.9
X 15 17 16.7 15.3 18.1 X -0.3 -1.1
X 12 19 X -3.3 0.9
X 17 20 X 1.7 1.9
Y 17 21 Y -1.5 -0.3
Y 19 20 Y 0.5 -1.3
Y 20 18 19.9 18.5 21.3 Y 1.5 -3.3
Y 17 22 Y -1.5 0.7
Y 22 23 Y 3.5 1.7

Average 16.9 19.7 18.3

Residuals Rank # P-% Z-value


-1.3 5 0.225 -0.755
0.7 12 0.575 0.189
-0.3 8 0.375 -0.319
-3.3 1 0.025 -1.960
1.7 17 0.825 0.935
-1.5 3 0.125 -1.150
0.5 11 0.525 0.063
1.5 16 0.775 0.755
-1.5 3 0.125 -1.150
3.5 20 0.975 1.960
-0.1 10 0.475 -0.063
0.9 14 0.675 0.454
-1.1 7 0.325 -0.454
0.9 14 0.675 0.454
1.9 19 0.925 1.440
-0.3 8 0.375 -0.319
-1.3 5 0.225 -0.755
-3.3 1 0.025 -1.960
0.7 12 0.575 0.189
1.7 17 0.825 0.935

The normal probability plot shows a reasonable linear relation. Thus, our
model assumption seems to fit and no obvious outliers are present.
11.2 Factorial designs with 3 or more factors
The most basic set-up of these factorial designs uses each factor at only two
levels (low and high).
And for a complete factorial test set-up all possible combinations between
these factors at the two levels have to be analyzed.
It is also necessary to perform at least one repeat test for each experiment in
order to evaluate all possible interactions.
A test without repeat measurements is possible when certain interactions are
assumed to be negligible.

General outline for a 23-factorial experiment:


Detergent-example:

Each factor has "high" and "low" values, e.g. Brand X is 'low" and Brand Y is
'high" and Warm is 'low' whereas Hot is 'high'.
The corner notations now indicate which factors are at a high or low level, e.g.
corner 'bc' represents the high values Hot and Liquid for the 2nd (b) and 3rd (c)
factor, whereas Brand X 1st (a) factor is at the low level.
So, how do we calculate the mean square values??
We could use the same approach as shown earlier.
For example:
a b c n'

∑∑∑∑ X
( )
a b c n' i jkl

sum of squares total; SST = ∑∑∑∑ X i jkl − X


2
i =1 j =1 k =1 l =1
with X =
i =1 j =1 k =1 l =1 a b c n'
b c n'

∑∑∑ X
) (
a i jkl

sum of squares factor A; SSA = b c n' ∑ X i ••• − X


2
j =1 k =1 l=1
with X i ••• =
i =1 b c n'

SSAB = c n' ∑∑ (X )
a b 2
sum of squares interaction AB; i j •• − X i ••• − X • j •• + X
i =1 j =1
sum of squares ABC;
a b c
(
SSABC = n' ∑∑∑ X i j k • − X i ••• − X • j •• − X ••k • + X i j•• − X i •k • − X • j k • + 5 X
i =1 j =1 k =1
) 2

Random error
( ) X i jkl
a b c n' n'
SSE = ∑∑∑∑ X i jkl − X ijk • with X ijk • = ∑
2
sum of squares error;
i =1 j =1 k =1 l =1 l =1 n'
Apparently these equations become quite cumbersome and your calculations
are prone to error.
However, there is a slightly easier way to calculate these sum of squares.
This is described as "Contrasts". The following table might come in handy.
In order to calculate the Contrast- or Effect-A we calculate the summation
over all experimental data according to the '+' and '-' signs shown in the
previous table.
Thus, data from the experiment labeled as (1) are counted negative, whereas
data from the experiment a are used with a positive value, etc..
This summation results in the so-called 'Contrast' for A.
Dividing the contrast by n' 2# of factors - 1 results in the so-called 'Effect'.

SS =
(Contrast )
2

The sum of squares is now: n' 2 # of factors

For our Laundry example the following examples may demonstrate this
process.

Effect A =
1
[Contrast ] = 1
[a + ab + ac + abc − (1) − b − c − bc] =
23-1 2 23-1 2
1
= 3-1
[18.4 + 18.8 + 20.7 + 21.1 + 21.7 + 21.9 + 23.4 + 24.6 −
2 2
18.8
14.8 − 15.0 − 17.8 − 18.2 − 19.9 − 20.3 − 22.6 − 23.2] = = 2.35
8
The Sum of Squares is now :

SSA =
(18.8)
2
= 22.09
23 2
The 3-parameter interaction ABC is calculated as follows:
Effect ABC = =
1
[a + b + c + abc − (1) − ab − ac − bc] =
23-1 2
1
= 3-1
[18.4 + 18.8 + 17.8 + 18.2 + 19.9 + 20.3 + 23.4 + 24.6 −
2 2
− 0.4
14.8 − 15.0 − 20.7 − 21.1 − 21.7 − 21.9 − 22.6 − 23.2] = = 0.05
8
The Sum of Squares is now :

SSABC =
(0.4)
2
= 0.01
23 2
The Sum of Square Total is given as:
( ) = 121.78
a b c n'
SS total = ∑∑∑∑ X
2
2
ijkl − a b c n' X
i =1 j =1 k =1 l =1

and the Sum of Squares Error is the difference between all individual
sum of squares and the SS total.
The following table summarizes the results:

Interaction effect df SS MS F P
Brand 1 22.090 22.090 140.25 0.000
Temperature 1 27.040 27.040 171.68 0.000
Type 1 67.240 67.240 426.92 0.000
Brand x Temperature 1 0.490 0.490 3.11 0.116
Brand x Type 1 3.610 3.610 22.92 0.001
Temperature x Type 1 0.040 0.040 0.25 0.628
Brand x Temperature x Type 1 0.010 0.010 0.06 0.807
Error 8 1.260 0.158
Total 15 121.780 121.780

It is obvious that all single effects and the Brand x Type interaction effect are
significant.
The larger the number of factors is, which are considered for the analysis, then
more experiments have to be preformed.
As a result of that quite often repeat experiments are not performed or a
fractional factorial design is being used.

First we evaluate the "no-repeat" process.


The following is a 25 factorial design:
The factorial analysis gives the following results:
(see excel file, "cake-analysis" for further details)
Interactions Sum of Squares df MS F P-value
A 1.4878 1 1.4878 3.03 0.102
B 0.9453 1 0.9453 1.92 0.186
C 4.8828 1 4.8828 9.93 0.007
D 0.0153 1 0.0153 0.03 0.862
E 0.0003 1 0.0003 0.00 0.980
AB 0.1378 1 0.1378 0.28 0.604
AC 0.0528 1 0.0528 0.11 0.748
AD 1.2403 1 1.2403 2.52 0.133
AE 0.0703 1 0.0703 0.14 0.711
BC 7.9003 1 7.9003 16.07 0.001
BD 0.1378 1 0.1378 0.28 0.604
BE 0.4753 1 0.4753 0.97 0.341
CD 0.6903 1 0.6903 1.40 0.254
CE 0.0253 1 0.0253 0.05 0.824
DE 38.5003 1 38.5003 78.31 0.000
ABC 0.4278 -0.4278 1 0.4278
ABD 0.9453 -0.9453 1 0.9453
ABE 0.1378 -0.1378 1 0.1378
ACD 0.0003 -0.0003 1 0.0003
ACE 1.2403 -1.2403 1 1.2403
ADE 11.4003 1 11.4003 23.19 0.000
BCD 0.0378 -0.0378 1 0.0378
BCE 0.0153 -0.0153 1 0.0153
BDE 0.5253 -0.5253 1 0.5253
CDE 1.2403 -1.2403 1 1.2403
ABCD 0.3003 -0.3003 1 0.3003
ABCE 0.2628 -0.2628 1 0.2628
ABDE 0.2628 -0.2628 1 0.2628
ACDE 0.025313 -0.0253 1 0.0253
BCDE 1.320313 -1.3203 1 1.3203
ABCDE 0.632813 -0.6328 1 0.6328

Error 7.3747 15 0.4916


Total 75.3372
In graphical form the results may be presented as follows:

11.3 Fractional factorial designs


The number of necessary experiments increases substantially for increasing
number of variables (factors). To reduce this number of experiments, we
'assume' certain types of interaction to be negligible and use subsets of the
complete factorial design.
We distinguish between different types of fractional designs:
Resolution III designs; where main effects are confounded with two-way
interactions (e.g. A confounded with BC)
Resolution IV designs; where main effects are not confounded with two-way
interactions (e.g. CD is confounded with BC)
Resolution V designs; where main effects and two-way interactions are not
confounded with each other, but higher order interactions are confounded
(e.g. ACD confounded with BCEF)
Example of a 24 factorial design:
A B C D AB AC AD BC BD CD ABC ABD ACD BCD ABCD
(1) - - - - + + + + + + - - - - +
a + - - - - - - + + + + + + - -
b - + - - - + + - - + + + - + -
c - - + - + - + - + - + - + + -
d - - - + + + - + - - - + + + -
ab + + - - + - - - - + - - + + +
ac + - + - - + - - + - - + - + +
ad + - - + - - + + - - + - - + +
bc - + + - - - + + - - - + + - +
bd - + - + - + - - + - + - + - +
cd - - + + + - - - - + + + - - +
abc + + + - + + - + - - + - - - -
abd + + - + + - + - + - - + - - -
acd + - + + - + + - - + - - + - -
bcd - + + + - - - + + + - - - + -
abcd + + + + + + + + + + + + + + +
For example, we decide to run only the experiments, where Effect ACD has a
positive sign in the previous table.
We can then see that the effects A & CD as well as C & AD or D & AC are
confounded (They have the same sign sequence and we are unable to
distinguish /analyze them individually)
Vice versa, if we had picked the positive experiments for column AC, then B
& ABC, D & ACD, as well as BC & ABCD are confounded. We should also
realize that the effect AC can no longer be analyzed at all.
If we decide to use only the positive experiments in the ABCD column, we
would then have the two-parameter interactions AB & CD, AC & BD, AD &
BC confounded.
In other cases you may use the same concept to reduce experimental variations
due to time or instrument limitations.
For example, if you can't run all experiment on the same day. You know that
results from the same date are very consistent, but repeat tests on different
dates show larger variations. You then perform groups (or blocks) of tests
according to the same principle as used before and you realize that for
example, that a possible significance of the ABCD interaction is caused by
your set-up.

In graphical form the block confounding is shown in the following graph:

The open and the closed circle experiments show the two blocks.
It is further possible to split the total number of experiments into ¼ blocks.
However, one has to be aware that more and more information is being lost in
that process.
E.g.
Block #1 Block #2 Block #3 Block #4
(1) a d ad
ab b abd bd
bc c bcd cd
ac abc acd abcd

For additional and more detailed information on factorial designs (e.g. factors at more than 2 levels) see:
O.L. Davis, The Design and Analysis of Industrial Experiments, Oliver and Boyd, London (1967)

You might also like