The sum of squared distances. The formulas presented are for a full factorial, two-factor model with factors A and B. These formulas can be extended to models with more than two factors. For information, see Montgomery1.
SS Total is the total variation in the model. SS (A) and SS (B) are the sum of the squared deviations of the estimated factor level means around the overall mean. SS Error is the sum of the squared residuals. It is also known as error within treatments. The calculations are:
Term | Description |
---|---|
a | number of levels in factor A |
b | number of levels in factor B |
n | total number of replicates |
mean of the ith level of factor A | |
overall mean of all observations | |
mean of the jth level of factor B | |
observation at the ith level of factor A, jth level of factor B, and the kth replicate | |
mean of the ith level of factor A and jth level of factor B | |
mean response for center points | |
mean response for factorial points | |
nF | number of factorial points |
Minitab breaks down the SS Regression or Treatments component of variance into sequential sums of squares for each factor. The sequential sums of squares depend on the order the factors or predictors are entered into the model. The sequential sum of squares is the unique portion of SS Regression explained by a factor, given any previously entered factors.
For example, if you have a model with three factors or predictors, X1, X2, and X3, the sequential sum of squares for X2 shows how much of the remaining variation X2 explains, given that X1 is already in the model. To obtain a different sequence of factors, repeat the analysis and enter the factors in a different order.
The adjusted sums of squares does not depend on the order that the terms enter the model. The adjusted sum of squares is the amount of variation explained by a term, given all other terms in the model, regardless of the order that the terms enter the model.
For example, if you have a model with three factors, X1, X2, and X3, the adjusted sum of squares for X2 shows how much of the remaining variation the term for X2 explains, given that the terms for X1 and X3 are also in the model.
The calculations for the adjusted sums of squares for three factors are:
where SSR(X3 | X1, X2) is the adjusted sum of squares for X3, given that X1 and X2 are in the model.
where SSR(X2, X3 | X1) is the adjusted sum of squares for X2 and X3, given that X1 is in the model.
You can extend these formulas if you have more than 3 factors in your model1.
For a full factorial design with factors A and B, and a blocking variable, the number of degrees of freedom associated with each sum of squares is:
For 2-level designs with center points, the degrees of freedom for curvature are 1.
Term | Description |
---|---|
a | number of levels in factor A |
b | number of levels in factor B |
c | number of blocks |
n | total number of observations |
ni | number of observations for ith factor level combination |
m | number of factor level combinations |
p | number of coefficients |
A test to determine whether the interaction and main effects are significant. The formula for the model terms is:
The degrees of freedom for the test are:
Larger values of F support rejecting the null hypothesis that there is not a significant effect.
For balanced split-plot designs, the F statistic for hard-to-change factors uses the MS for whole plot error in the denominator. For other split-plot designs, Minitab uses a linear combination of the WP Error and the SP Error to construct a denominator that is based on the expected mean squares.
The p-value is a probability that is calculated from an F-distribution with the degrees of freedom (DF) as follows:
1 − P(F ≤ fj)
Term | Description |
---|---|
P(F ≤ f) | cumulative distribution function for the F-distribution |
f | f-statistic for the test |
where n = number of observations and m = number of distinct x-level combinations
Large F-values and small p-values suggest that the model is inadequate.
1 − P(F ≤ fj)
Term | Description |
---|---|
P(F ≤ fj) | cumulative distribution function for the F-distribution |
fj | f-statistic for the test |