See also: What is an F-Test?ANOVA test or a regression analysis to find out if the means between two populations are significantly different. It’s similar to a T statistic from a T-Test; A-T test will tell you if a single variable is statistically significant and an F test will tell you if a group of variables are jointly significant.
What is “Statistically Significant“?
Simply put, if you have significant result, it means that your results likely did not happen by chance. If you don’t have statistically significant results, you throw your test data out (as it doesn’t show anything!); in other words, you can’t reject the null hypothesis.
Using The F Statistic.
You can use the F statistic when deciding to support or reject the null hypothesis. In your F test results, you’ll have both an F value and an F critical value.
- The F critical value is also called the F statistic.
- The value you calculate from your data is called the F value (without the “critical” part).
In general, if your calculated F value in a test is larger than your F statistic, you can reject the null hypothesis. However, the statistic is only one measure of significance in an F Test. You should also consider the p value. The p value is determined by the F statistic and is the probability your results could have happened by chance.
Back to Top
The F statistic must be used in combination with the p value when you are deciding if your overall results are significant. Why? If you have a significant result, it doesn’t mean that all your variables are significant. The statistic is just comparing the joint effect of all the variables together.
- If the p value is less than the alpha level, go to Step 2 (otherwise your results are not significant and you cannot reject the null hypothesis). A common alpha level for tests is 0.05.
- Study the individual p values to find out which of the individual variables are statistically significant.
The F value in one way ANOVA is a tool to help you answer the question “Is the variance between the means of two populations significantly different?” The F value in the ANOVA test also determines the P value; The P value is the probability of getting a result at least as extreme as the one that was actually observed, given that the null hypothesis is true.
The p value is a probability, while the f ratio is a test statistic, calculated as:
When Do I Reject the Null Hypothesis?
Reject the null when your p value is smaller than your alpha level. You should not reject the null if your critical f value is smaller than your F Value, unless you also have a small p-value.
Where this could get confusing is where one of these values seems to indicate that you should reject the null hypothesis and one of the values indicates you should not. For example, let’s say your One Way ANOVA has a p value of 0.68 and an alpha level of 0.05. As the p value is large, you should not reject the null hypothesis. However, your f value is 4.0 with an f critical value of 3.2. Should you now reject the null hypothesis? The answer is NO.
The F value should always be used along with the p value in deciding whether your results are significant enough to reject the null hypothesis. If you get a large f value (one that is bigger than the F critical value found in a table), it means something is significant, while a small p value means all your results are significant. The F statistic just compares the joint effect of all the variables together. To put it simply, reject the null hypothesis only if your alpha level is larger than your p value.
Caution: If you are running an F Test in Excel, make sure your variance 1 is smaller than variance 2. This “quirk” can give you an incorrect f ratio if you put the variances in the wrong place. See the bottom of this article for an example: F Test Two Sample Variances in Excel.
Back to Top
The F value in regression is the result of a test where the null hypothesis is that all of the regression coefficients are equal to zero. In other words, the model has no predictive capability. Basically, the f-test compares your model with zero predictor variables (the intercept only model), and decides whether your added coefficients improved the model. If you get a significant result, then whatever coefficients you included in your model improved the model’s fit.
Read your p-value first. If the p-value is small (less than your alpha level), you can accept the null hypothesis. Only then should you consider the f-value. If you fail to reject the null, discard the f-value result.
Many authors recommend ignoring the P values for individual regression coefficients if the overall F ratio is not statistically significant. This is because of the multiple testing problem. In other words, your p-value and f-value should both be statistically significant in order to correctly interpret the results.
If you want to know whether your regression F-value is significant, you’ll need to find the critical value in the f-table. For example, let’s say you had 3 regression degrees of freedom (df1) and 120 residual degrees of freedom (df2). An F statistic of at least 3.95 is needed to reject the null hypothesis at an alpha level of 0.1. At this level, you stand a 1% chance of being wrong (Archdeacon, 1994, p.168). For more details on how to do this, see: F Test. F Values will range from 0 to an arbitrarily large number.
Back to Top
The F Distribution is a probability distribution of the F Statistic. In other words, it’s a distribution of all possible values of the f statistic.
The distribution is an asymmetric distribution usually used for ANOVA . It has a minimum value of zero; there is no maximum value. The distribution’s peak happens just to the right of zero and the higher the f-value after that point, the lower the curve. The F distribution is actually a collection of distribution curves. The F distribution is related to chi-square, because the f distribution is the ratio of two chi-square distributions with degrees of freedom ν1 and ν2 (note: each chi-square is first been divided by its degrees of freedom). Each curve depends on the degrees of freedom in the numerator (dfn) and the denominator (dfd). These depend upon your sample characteristics.
For example, in a simple one-way ANOVA between-groups,
- Dfn = a – 1
- dfd = N – a
- a = the number of groups
- n = the total number of subjects in the experiment
The degrees of freedom in the denominator (dfd) is also referred to as the degrees of freedom error (dfe).
The F Distribution is also called the Snedecor’s F, Fisher’s F or the Fisher–Snedecor distribution.
Back to Top
There are two types of main problem you’ll encounter with the F-Distribution you might be asked to find the area under a F curve given numerator degrees of freedom (ndf), denominator degrees of freedom (ddf), and a certain range (for example, P( 1 ≤ X ≤ 2 ), or you might be asked to find the F value with area to the left, a certain ndf and ddf (useful for finding critical values for hypotheses tests).
F Distribution on TI 89: Steps
Sample problem: find the area under a F curve with numerator degrees of freedom (ndf) 4 and
denominator degrees of freedom (ddf) 10 for For, P( 1≤ X ≤ 2 ):
Step 1: Press APPS.
Step 2:Press ENTER twice to get to the list entry screen.
Step 3: Press F5 for “F5-Distr.”
Step 4: Scroll down to “A:F Cdf” and press ENTER.
Step 5: Enter 1 in the box for “Lower Value,” then press the down arrow key.
Step 6: Enter 2 in the box for “Upper Value,” then press the down arrow key.
Step 7: Enter 4 in the “Num df” box, then press the down arrow key.
Step 8: Enter 5 in the “Den df” box.
Step 9: Press ENTER. The calculator will return .281 as the answer.
Sample problem: to find the F value with area to the left, with ndf = 5, ddf = 8, and an area of .99:
Step 1: Press APPS.
Step 2: Press ENTER twice to get to the list entry screen.
Step 3: Press F5 for “F5-Distr.”
Step 4: Press 2 for “Inverse.”
Step 5: Press 4 for “Inverse F…,” then press ENTER.
Step 6: Enter .99 in the “Area” box, then press the down arrow key.
Step 7: Enter 5 in the “Num df,” box, then press the down arrow key.
Step 8: Enter 8 in the “Den df.” box, then press ENTER. This returns the answer (63183).
Tip: For P( X ≥ 1 ), enter 1 in the box for Lower Value and 10 ^ 99 in the box for
Upper Value, and for For P( X ≤ 1 ), enter 0 in the box for Lower Value, then enter 1 in the box for Upper
The F Table is a collection of tables that give you the probability for a certain alpha level. The F Table is actually a collection of tables, for four alpha levels: .10. .5, .025 and .01.
The three f tables you can find on this site are for alpha levels of .10, .0 and .01. When using the F dist. table, always put the numerator degrees of freedom first; if you switch the numerator and denominator around, you’ll get a different result. The table gives you the area in the right tail. Instead of a table, you can use a calculator — which will give you more accurate results.
What is the F Statistic Table Used for?
When you have found the F value, you can compare it with an f critical value in the table. If your observed value of F is larger than the value in the F table, then you can reject the null hypothesis with 95 percent confidence that the variance between your two populations isn’t due to random chance.
How to use the F Statistic Table
The F Statistic Table is actually a collection of tables. Which specific table you use will depend on which alpha level you use. For example, if you have an alpha level of .05, then your right tail area is .05 (5 percent), and you’ll look up the f critical value in the alpha level = .05 table. The rows in the F Distribution Table represent denominator degrees of freedom and the columns represent numerator degrees of freedom.
For example, to determine the .10 f critical value for an F distribution with 6 and 6 degrees of freedom, look in the 6 column (numerator) and the 6 row (denominator) of the F Table for alpha=.10. F(.10, 6, 6) = 3.05455.
Why Use the F Statistic Table? Why not just use a calculator?
A calculator will certainly give you a fast answer. But with many scenarios in statistics, you will look at a range of possibilities and a table is much better for visualizing a large number of probabilities at the same time.
Back to Top
Archdeacon, T. (1994). Correlation and Regression Analysis: A Historian’s Guide. Univ of Wisconsin Press.