This chapter introduces a new probability density function, the F distribution. This distribution is used for many applications including ANOVA and for testing equality across multiple means. We begin with the F distribution and the test of hypothesis of differences in variances. It is often desirable to compare two variances rather than two averages. For instance, college administrators would like two college professors grading exams to have the same variation in their grading. In order for a lid to fit a container, the variation in the lid and the container should be approximately the same. A supermarket might be interested in the variability of check-out times for two checkers. In finance, the variance is a measure of risk and thus an interesting question would be to test the hypothesis that two different investment portfolios have the same variance, the volatility.

In order to perform a *F* test of two variances, it is important that the following are true:

- The populations from which the two samples are drawn are approximately normally distributed.
- The two populations are independent of each other.

Unlike most other hypothesis tests in this book, the *F* test for equality of two variances is very sensitive to deviations from normality. If the two distributions are not normal, or close, the test can give a biased result for the test statistic.

Suppose we sample randomly from two independent normal populations. Let ${\sigma}_{1}^{2}$ and ${\sigma}_{2}^{2}$ be the unknown population variances and ${s}_{1}^{2}$ and ${s}_{2}^{2}$ be the sample variances. Let the
sample sizes be *n*_{1} and *n*_{2}. Since we are interested in comparing the two sample variances, we use the *F* ratio:

$F=\frac{\left[\frac{{{s}_{1}}^{2}}{{{\sigma}_{1}}^{2}}\right]}{\left[\frac{{{s}_{2}}^{2}}{{{\sigma}_{2}}^{2}}\right]}$

*F* has the distribution *F* ~ *F*(*n*_{1} – 1, *n*_{2} – 1)

where *n*_{1} – 1 are the degrees of freedom for the numerator and *n*_{2} – 1 are the degrees of freedom for the denominator.

If the null hypothesis is ${\sigma}_{1}^{2}={\sigma}_{2}^{2}$, then the *F* Ratio, test statistic, becomes ${F}_{c}=\frac{\left[\frac{{{s}_{1}}^{2}}{{{\sigma}_{1}}^{2}}\right]}{\left[\frac{{{s}_{2}}^{2}}{{{\sigma}_{2}}^{2}}\right]}=\frac{{{s}_{1}}^{2}}{{{s}_{2}}^{2}}$

The various forms of the hypotheses tested are:

Two-Tailed Test | One-Tailed Test | One-Tailed Test |
---|---|---|

H_{0}: σ_{1}^{2} = σ_{2}^{2} |
H_{0}: σ_{1}^{2} ≤ σ_{2}^{2} |
H_{0}: σ_{1}^{2} ≥ σ_{2}^{2} |

H_{1}: σ_{1}^{2} ≠ σ_{2}^{2} |
H_{1}: σ_{1}^{2} > σ_{2}^{2} |
H_{1}: σ_{1}^{2} < σ_{2}^{2} |

A more general form of the null and alternative hypothesis for a two tailed test would be :

Where if δ_{0} = 1 it is a simple test of the hypothesis that the two variances are equal. This form of the hypothesis does have the benefit of allowing for tests that are more than for simple differences and can accommodate tests for specific differences as we did for differences in means and differences in proportions. This form of the hypothesis also shows the relationship between the F distribution and the χ^{2} : the F is a ratio of two chi squared distributions a distribution we saw in the last chapter. This is helpful in determining the degrees of freedom of the resultant F distribution.

If the two populations have equal variances, then ${s}_{1}^{2}$ and ${s}_{2}^{2}$ are close in value and the test statistic, ${F}_{c}=\frac{{{s}_{1}}^{2}}{{{s}_{2}}^{2}}$ is close to one. But if the two population variances are very different, ${s}_{1}^{2}$ and ${s}_{2}^{2}$ tend to be very different, too. Choosing ${s}_{1}^{2}$ as the larger sample variance causes the ratio $\frac{{{s}_{1}}^{2}}{{{s}_{2}}^{2}}$ to be greater than one. If ${s}_{1}^{2}$ and ${s}_{2}^{2}$ are far apart, then ${F}_{c}=\frac{{{s}_{1}}^{2}}{{{s}_{2}}^{2}}$ is a large number.

Therefore, if *F* is close to one, the evidence favors the null hypothesis (the two population variances are equal). But if *F* is much larger than one, then the evidence is against the null hypothesis. In essence, we are asking if the calculated F statistic, test statistic, is significantly different from one.

To determine the critical points we have to find F_{α},_{df1},_{df2}. See Appendix A for the F table. This F table has values for various levels of significance from 0.1 to 0.001 designated as "p" in the first column. To find the critical value choose the desired significance level and follow down and across to find the critical value at the intersection of the two different degrees of freedom. The F distribution has two different degrees of freedom, one associated with the numerator, _{df1}, and one associated with the denominator, _{df2} and to complicate matters the F distribution is not symmetrical and changes the degree of skewness as the degrees of freedom change. The degrees of freedom in the numerator is n_{1}-1, where n_{1} is the sample size for group 1, and the degrees of freedom in the denominator is n_{2}-1, where n_{2} is the sample size for group 2. F_{α},_{df1},_{df2} will give the critical value on the **upper** end of the F distribution.

To find the critical value for the **lower** end of the distribution, reverse the degrees of freedom and divide the F-value from the table into one.

- Upper tail critical value : F
_{α},_{df1},_{df2} - Lower tail critical value : 1/F
_{α},_{df2},_{df1}

When the calculated value of F is between the critical values, not in the tail, we cannot reject the null hypothesis that the two variances came from a population with the same variance. If the calculated F-value is in either tail we cannot accept the null hypothesis just as we have been doing for all of the previous tests of hypothesis.

An alternative way of finding the critical values of the F distribution makes the use of the F-table easier. We note in the F-table that all the values of F are greater than one therefore the critical F value for the left hand tail will always be less than one because to find the critical value on the left tail we divide an F value into the number one as shown above. We also note that if the sample variance in the numerator of the test statistic is larger than the sample variance in the denominator, the resulting F value will be greater than one. The shorthand method for this test is thus to be sure that the larger of the two sample variances is placed in the numerator to calculate the test statistic. This will mean that only the right hand tail critical value will have to be found in the F-table.

### Example 12.1

Two college instructors are interested in whether or not there is any variation in the way they grade math exams. They each grade the same set of 10 exams. The first instructor's grades have a variance of 52.3. The second instructor's grades have a variance of 89.9. Test the claim that the first instructor's variance is smaller. (In most colleges, it is desirable for the variances of exam grades to be nearly the same among instructors.) The level of significance is 10%.

The New York Choral Society divides male singers up into four categories from highest voices to lowest: Tenor1, Tenor2, Bass1, Bass2. In the table are heights of the men in the Tenor1 and Bass2 groups. One suspects that taller men will have lower voices, and that the variance of height may go up with the lower voices as well. Do we have good evidence that the variance of the heights of singers in each of these two groups (Tenor1 and Bass2) are different?

Tenor1 | Bass 2 | Tenor 1 | Bass 2 | Tenor 1 | Bass 2 |
---|---|---|---|---|---|

69 | 72 | 67 | 72 | 68 | 67 |

72 | 75 | 70 | 74 | 67 | 70 |

71 | 67 | 65 | 70 | 64 | 70 |

66 | 75 | 72 | 66 | 69 | |

76 | 74 | 70 | 68 | 72 | |

74 | 72 | 68 | 75 | 71 | |

71 | 72 | 64 | 68 | 74 | |

66 | 74 | 73 | 70 | 75 | |

68 | 72 | 66 | 72 |