![]() ![]() The second is that there is a difference, and it is called the alternative, and is denoted H 1 or H a. The first is the null hypothesis that there is no difference (hence null). Formally, two hypotheses are needed for completeness. There are two sets of details: first, formally writing hypotheses, and second, using the F-distribution tables so that you can tell if your F-score is close to one or not. The basic method must be fleshed out with some details if you are going to use this test at work. If the F-score is far from one, then conclude that the populations probably have different variances. If the F-score is close to one, conclude that your hypothesis is correct and that the samples do come from populations with equal variances. Compute the F-score by finding the ratio of the sample variances. Hypothesize that the samples come from populations with the same variance. ![]() If you compute the F-score, and it is close to one, you accept your hypothesis that the samples come from populations with the same variance. Because the F-statistic is the ratio of two sample variances, when the two sample variances are close to equal, the F-score is close to one. Obviously, if the two variances are very close to being equal the two samples could easily have come from populations with equal variances. You would have two samples (one of size n 1 and one of size n 2) and the sample variance from each. A simple test: Do these two samples come from populations with the same variance?īecause the F-distribution is generated by drawing two samples from the same normal population, it can be used to test the hypothesis that two samples come from populations with the same variance. The second is one-way analysis of variance (ANOVA), which uses the F-distribution to test to see if three or more samples come from populations with the same mean. The first is a very simple test to see if two samples come from populations with the same variance. There are two uses of the F-distribution that will be discussed in this chapter. Since F goes from zero to very large, with most of the values around one, it is obviously not symmetric there is a long tail to the right, and a steep descent to zero on the left. It is equally possible for s 2 2 to be a lot larger than s 1 2, and then F would be very close to zero. If s 1 2 is a lot larger than s 2 2, F can be quite large. ![]() Thinking about ratios requires some care. All of the F-scores will be positive since variances are always positive - the numerator in the formula is the sum of squares, so it will be positive, the denominator is the sample size minus one, which will also be positive. If s 1 2 and s 2 2 come from samples from the same population, then if many pairs of samples were taken and F-scores computed, most of those F-scores would be close to one. Think about the shape that the F-distribution will have. Because we know that sampling distributions of the ratio of variances follow a known distribution, we can conduct hypothesis tests using the ratio of variances. Not surprisingly, over the intervening years, statisticians have found that the ratio of sample variances collected in a number of different ways follow this same distribution, the F-distribution. Years ago, statisticians discovered that when pairs of samples are taken from a normal population, the ratios of the variances of the samples in each pair will always follow the same distribution.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |