What Is an ANOVA Table and How Do You Read It?

An ANOVA table is a structured summary that shows whether the differences between group averages in a dataset are statistically meaningful or just due to random chance. ANOVA stands for “analysis of variance,” and the table breaks down the total variability in your data into distinct sources, then tests whether the variation between groups is large enough to conclude that something real is going on. If you’ve ever compared test scores across three classrooms, plant growth under different fertilizers, or customer satisfaction across multiple stores, the ANOVA table is the standard tool for summarizing that comparison.

What Each Column Means

Every ANOVA table has the same core columns, regardless of the software that produced it. Understanding these six columns is the key to reading any ANOVA output.

  • Source of Variation: The rows of the table, identifying where the variability comes from. In a one-way ANOVA, you’ll see “Between Groups” (the differences caused by your grouping variable), “Within Groups” or “Error” (the natural variation inside each group), and “Total.”
  • Sum of Squares (SS): A measure of total variability attributed to each source. The between-groups SS captures how far each group’s average is from the overall average. The within-groups SS captures how far individual data points are from their own group’s average. These two always add up to the total SS.
  • Degrees of Freedom (df): The number of independent pieces of information used in each calculation. Between-groups df equals the number of groups minus one. Within-groups df equals the total number of observations minus the number of groups. These also add up to the total df.
  • Mean Square (MS): The sum of squares divided by its degrees of freedom. This converts raw variability into an average-per-unit measure so the between and within sources can be fairly compared.
  • F-statistic: The ratio of the between-groups mean square to the within-groups mean square. If the groups are truly no different from each other, you’d expect this ratio to hover around 1. The further it climbs above 1, the stronger the evidence that at least one group differs.
  • P-value: The probability of getting an F-statistic as large as the one you observed if there were genuinely no difference between the groups. A small p-value (commonly below 0.05) means the group differences are unlikely to be random.

How the Calculations Work

The logic flows from left to right across the table. First, the total variability in the data gets split into two pieces: variation between groups and variation within groups. Then each piece is standardized, and the two standardized values are compared.

Say you’re comparing the exam scores of students taught by three different methods, with a total of 30 students. The between-groups sum of squares measures how much the average score for each method deviates from the grand average of all 30 scores. The within-groups sum of squares measures how much individual students deviate from their own method’s average. The between-groups df is 3 minus 1, which gives you 2. The within-groups df is 30 minus 3, which gives you 27.

Dividing each sum of squares by its degrees of freedom produces the mean squares. The between-groups mean square is then divided by the within-groups mean square to get the F-statistic. If teaching method truly affects scores, the between-groups mean square will be substantially larger than the within-groups mean square, pushing the F-statistic well above 1. The p-value then tells you exactly how surprising that F-value would be under the assumption that teaching method doesn’t matter at all.

Reading the F-Statistic and P-Value

The F-statistic is the single number that answers the core question: is the variation between groups bigger than what you’d expect from random noise alone? An F-value near 1 suggests your groups aren’t meaningfully different. An F-value of, say, 5 or 10 suggests the group differences are several times larger than the background noise in the data.

The p-value translates that F-statistic into a probability. It answers: “If there were truly no difference between these groups, what’s the chance I’d see an F-value this large just by luck?” When the p-value drops below your chosen threshold (0.05 is the most common), you reject the idea that all group averages are equal. But a significant result only tells you that at least one group differs. It doesn’t tell you which group, or which pairs of groups, are different from each other.

What the Table Doesn’t Tell You

A significant F-test tells you something differs, but not where the difference lies. If you’re comparing four fertilizers and get a significant result, you know at least one fertilizer performs differently, but you don’t know if it’s fertilizer A versus B, A versus C, or some other combination. This is where post-hoc tests come in. A post-hoc test (such as Tukey’s HSD or Bonferroni correction) is used only after you find a statistically significant ANOVA result, and it systematically compares every pair of groups to pinpoint exactly where the differences are.

The ANOVA table also doesn’t tell you how large the effect is. A tiny difference can be statistically significant with a big enough sample, so researchers often report an effect size alongside the table. The most common one is eta-squared, which equals the between-groups sum of squares divided by the total sum of squares. It tells you the proportion of all variability in the data that’s explained by group membership. An eta-squared of 0.10, for example, means the grouping variable accounts for 10% of the total variation. Partial eta-squared is a related measure that can be calculated directly from the F-statistic and degrees of freedom, making it useful when you only have the ANOVA output rather than the raw data.

One-Way vs. Two-Way ANOVA Tables

A one-way ANOVA table has the simplest structure: one row for between-groups variation, one for within-groups (error), and one for the total. You’re testing a single factor, like whether three diets produce different weight loss.

A two-way ANOVA table adds rows because you’re testing two factors at once, plus their interaction. If you’re studying the effect of both diet type and exercise level on weight loss, the table breaks the between-groups variation into three separate lines: one for diet, one for exercise, and one for the diet-by-exercise interaction. The sums of squares and degrees of freedom for these three lines add up to the total between-groups values, just as they would in a one-way layout. Each line gets its own F-statistic and p-value.

The interaction row is often the most interesting part of a two-way table. Its null hypothesis is that the two factors have a purely additive relationship, meaning the effect of diet is the same regardless of exercise level. A significant interaction tells you the factors don’t work independently: perhaps one diet only works well when paired with high exercise. The interaction degrees of freedom follow a simple rule. For a design with k levels of one factor and m levels of the other, interaction df equals (k minus 1) times (m minus 1).

Assumptions Behind the Table

The numbers in an ANOVA table are only trustworthy when three conditions hold. First, the observations need to be independent of each other, meaning one person’s score doesn’t influence another’s. Second, the data within each group should be roughly normally distributed, though ANOVA is fairly forgiving of mild departures from normality, especially with larger samples. Third, the groups should have similar amounts of variability, a condition called homogeneity of variance. If one group’s scores are spread across a huge range while another group’s scores are tightly clustered, the F-test can give misleading results.

When these assumptions are seriously violated, alternatives exist. Non-parametric tests like the Kruskal-Wallis test don’t require normality, and corrections like Welch’s ANOVA handle unequal variances. But for most well-designed studies with reasonable sample sizes, the standard ANOVA table remains reliable and straightforward to interpret.