What Is an ANOVA Table? Columns and Values Explained

An ANOVA table is a structured summary that breaks down the total variation in a dataset into meaningful parts: variation caused by the groups or treatments you’re comparing, and variation caused by random chance. ANOVA stands for “analysis of variance,” and the table is the standard way to display the results of this test. If you’re taking a statistics course or reading a research paper, understanding this table is essential for interpreting whether group differences are real or just noise in the data.

What an ANOVA Table Actually Does

At its core, an ANOVA table answers one question: do the means of two or more groups differ more than you’d expect by chance alone? It does this by splitting the total variability in your data into two buckets. One bucket captures how much the group averages differ from each other (the “between groups” variation). The other captures how much individual data points within each group scatter around their own group’s average (the “within groups” or “error” variation).

If the between-groups variation is large relative to the within-groups variation, the table gives you evidence that something systematic is going on, not just random scatter. The table organizes all of this into a few columns that build on each other, culminating in two key numbers: the F-statistic and the p-value.

The Standard Columns

A typical ANOVA table has six columns. Here’s what each one means and why it matters.

  • Source: Identifies where the variation comes from. You’ll usually see rows labeled “Between Groups” (or “Treatment” or “Factor”), “Within Groups” (or “Error”), and “Total.”
  • SS (Sum of Squares): A measure of total variation for each source. Larger numbers mean more spread in the data. The SS values are the raw building blocks of the entire table.
  • df (Degrees of Freedom): The number of independent pieces of information used to calculate each sum of squares. Think of it as an adjustment for sample size and the number of groups.
  • MS (Mean Square): The sum of squares divided by its degrees of freedom. This converts raw variation into an average variation per unit, which makes fair comparisons possible.
  • F: The F-statistic, calculated by dividing the mean square between groups by the mean square within groups. This single ratio tells you how large the group differences are relative to the background noise.
  • p: The probability of seeing an F-value this large (or larger) if the groups actually had identical means. A small p-value suggests the group differences are real.

How Sum of Squares Works

The sum of squares (SS) is where all the math starts. The total sum of squares measures how far every individual observation falls from the overall average of the entire dataset. It captures all the variability present, regardless of which group each observation belongs to.

That total gets partitioned into two pieces. The “between groups” sum of squares looks at how far each group’s average is from the overall average, weighted by the number of observations in each group. If your groups have very different averages, this number will be large. The “within groups” sum of squares looks at how far individual observations fall from their own group’s average. This reflects the natural scatter that exists even within a single group, the kind of variation that has nothing to do with your treatment or factor.

These two pieces add up to the total: SS Between + SS Within = SS Total. That additive relationship is a defining feature of the ANOVA table, and it’s what makes the decomposition useful.

Degrees of Freedom

Degrees of freedom act as a correction factor. For the between-groups row, the degrees of freedom equal the number of groups minus one. If you’re comparing four treatments, that’s 3 degrees of freedom. For the within-groups row, the degrees of freedom equal the total number of observations minus the number of groups. So if you have 30 total observations across those four groups, the within-groups df is 26. The total degrees of freedom is simply the total number of observations minus one (29 in this example).

These numbers matter because raw sums of squares don’t account for how many data points contributed to them. A group with 100 observations will naturally produce a larger sum of squares than a group with 10, even if variability is the same. Dividing by degrees of freedom levels the playing field.

Mean Square and the F-Ratio

The mean square for each row is simply its sum of squares divided by its degrees of freedom. This gives you a per-unit measure of variation that can be compared fairly across rows.

The F-statistic is the ratio of the between-groups mean square to the within-groups mean square. If the groups are truly no different from one another, both mean squares should estimate roughly the same underlying variability, and the F-ratio should hover around 1. The further F climbs above 1, the stronger the evidence that group membership actually matters.

An F-value of, say, 0.8 suggests the groups vary less than you’d expect from chance alone. An F-value of 12 suggests the between-group differences are 12 times larger than what random within-group scatter would predict, a strong signal.

Interpreting the P-Value

The p-value translates the F-statistic into a probability. It tells you how likely it would be to observe your F-value (or something more extreme) if there were truly no differences between groups. A p-value of 0.03, for instance, means there’s only a 3% chance of getting results this extreme under the assumption that all group means are equal.

Researchers typically use a threshold of 0.05. Below that, they reject the null hypothesis and conclude the groups differ. Below 0.01, the evidence is considered quite strong. A p-value of 0.20 or higher offers little reason to believe the groups are meaningfully different.

One important caveat: the F-test assumes the data within each group follow a roughly normal distribution and that the variability is similar across groups. When those assumptions are violated, a low p-value is still generally trustworthy, but borderline values (say, around 0.05 to 0.10) should be interpreted more cautiously.

Effect Size From the Table

A p-value tells you whether group differences exist, but not how large they are. For that, you can calculate an effect size called eta-squared directly from the ANOVA table. The formula is straightforward: divide the between-groups sum of squares by the total sum of squares.

The result is a proportion between 0 and 1 that tells you what percentage of the total variability in your data is explained by the grouping factor. An eta-squared of 0.34 means the factor explains 34.1% of the variability in the outcome, which is a substantial effect. Values around 0.01 are considered small, 0.06 medium, and 0.14 or above large. In academic papers formatted to APA standards, you’ll often see eta-squared reported alongside the F-statistic.

Two-Way ANOVA Tables

The table described above applies to a one-way ANOVA, which tests the effect of a single factor. A two-way ANOVA table adds rows because it tests two factors simultaneously, plus their interaction.

For example, if you’re studying how both diet type and exercise level affect weight loss, a two-way table would have separate rows for diet (Factor A), exercise (Factor B), the interaction between diet and exercise (A × B), within-groups error, and the total. Each row gets its own SS, df, MS, F-statistic, and p-value. The interaction row is especially important because it tells you whether the effect of one factor depends on the level of the other. A significant interaction between diet and exercise would mean certain diet-exercise combinations produce effects you wouldn’t predict just by adding their individual effects together.

The same logic of decomposing total variation still applies. All the sum of squares rows still add up to the total, and each F-ratio still compares a source of variation against the error term.

Reading ANOVA Output in Software

Most people encounter ANOVA tables through software like Excel, SPSS, R, or Python rather than building them by hand. The labels vary slightly between programs. Excel’s Data Analysis tool labels the rows “Between Groups” and “Within Groups.” SPSS often uses “Between Groups” and “Error.” R’s output might show the factor name directly (like “Diet”) and label the error row “Residuals.”

Despite the labeling differences, the columns are always the same: source, sum of squares, degrees of freedom, mean square, F-statistic, and p-value. Some software adds a column for eta-squared or other effect sizes automatically. Others require you to calculate it from the SS values in the table. Regardless of the tool, the interpretation process is identical: find the F-value, check the p-value, and calculate the effect size if it isn’t provided.