A probability plot is a graph that shows whether your data follows a particular statistical distribution, such as a bell curve. You plot your actual data points against where they’d theoretically fall if they perfectly matched that distribution. If the points form a roughly straight line, your data fits the distribution well. If the points curve or bend away from the line, your data doesn’t match.
How a Probability Plot Works
The basic idea is a visual comparison. You take your dataset, sort every value from smallest to largest, and then ask: if this data truly came from a normal distribution (or whichever distribution you’re testing), where would each value be expected to land? You plot your actual values on one axis and the expected theoretical values on the other. Each data point becomes a dot on the graph.
When the dots line up along a diagonal, you’re looking at a good fit. The data behaves the way that distribution predicts. When the dots curve, cluster, or fan out from the line, something about your data doesn’t match the assumed distribution. The strength of that straight-line pattern can even be measured with a correlation coefficient, giving you a number to go with the visual.
What the Plot Actually Calculates
Behind the scenes, the plot assigns each sorted data point a rank from 1 (smallest) to n (largest), where n is your sample size. Those ranks get converted into cumulative probabilities, essentially answering the question “what percentage of data falls at or below this point?” This is called the empirical distribution function, and it’s your data’s way of describing itself.
Those cumulative probabilities are then fed through the inverse of the theoretical distribution you’re testing against. For a normal probability plot, this converts each probability into the corresponding value you’d expect from a perfect bell curve. The result is a set of “expected” values paired with your actual sorted data. Plot one against the other, and you have your probability plot.
Several formulas exist for converting ranks into probabilities. The most common ones carry names like Blom, Tukey, and Filliben, and they differ slightly in how they handle the edges of your data. Most statistical software picks one by default (often Filliben’s or a variant called RankIt), and for practical purposes the differences between them are small.
P-P Plots vs. Q-Q Plots
You’ll often see two variations mentioned together. A P-P plot (probability-probability plot) compares cumulative probabilities directly. It converts your data into z-scores, then plots your actual z-scores against the z-scores you’d expect from a normal distribution. A Q-Q plot (quantile-quantile plot) does something similar but compares quantiles, the values that split your data into equal portions, rather than plotting every individual score. Both produce the same straight-line-means-good-fit interpretation. Q-Q plots tend to be more sensitive to differences in the tails of a distribution, making them slightly better at catching outliers and heavy-tailed behavior.
Reading the Shape of the Plot
The real value of a probability plot is what the deviations tell you. A perfectly straight line means the data matches the distribution. But the specific ways points deviate from the line reveal what’s actually going on with your data.
If the points curve upward at both ends (forming an S-shape that bows above the line on the right and below it on the left), your data has heavier tails than the theoretical distribution. This means more extreme values than expected, both high and low. The reverse S-shape, where points fall below the line at the right end and above it at the left, indicates lighter tails than expected, meaning your data is more tightly clustered around the center.
A curve that bends consistently in one direction signals skewness. When your data is skewed right (a long tail stretching toward higher values), the points will curve upward away from the line at the high end. Left-skewed data does the opposite. A single point sitting far from the line while the rest behave nicely is a classic sign of an outlier. If the lower and upper extremes of the plot stay close to the fitted line, you can be reasonably confident there are no significant outliers relative to the distribution you’re testing.
Why It Matters for Statistical Tests
Many common statistical procedures, including t-tests, correlation, regression, and analysis of variance, assume your data comes from a normal distribution. If that assumption is wrong, the results of those tests can be misleading. Probability plots give you a quick, intuitive way to check before running your analysis.
Formal normality tests (like the Shapiro-Wilk test) give you a yes-or-no answer based on a p-value, but they have a well-known limitation: with large sample sizes, they flag trivially small departures from normality as “significant,” and with small samples, they lack the power to detect real problems. A probability plot shows you the shape and severity of any deviation, letting you judge whether it’s large enough to actually matter for your analysis. True perfect normality is considered a myth in real-world data. The practical question is whether your data is close enough.
Sample Size and Reliability
Probability plots work with any sample size, but they become harder to interpret with very small datasets. With only 10 or 15 data points, random variation alone can produce plots that look curved even when the data genuinely comes from the target distribution. You’ll see more scatter around the line simply because you have fewer points to establish a pattern. As sample size grows, the plot becomes more stable and deviations from the line become more meaningful.
The general guidance is to treat probability plots from small samples as suggestive rather than definitive. If you have 30 or more observations, the visual pattern becomes considerably more reliable. With hundreds of data points, even subtle departures from a distribution become visible.
Testing Distributions Beyond the Bell Curve
Normal probability plots are the most common type, but the technique works for any distribution you can define mathematically. You can construct probability plots for exponential, Weibull, lognormal, or other distributions simply by swapping in the appropriate theoretical function when calculating the expected values. A Weibull probability plot, for instance, is widely used in engineering to model failure times and product lifespans. The interpretation stays the same: straight line means the distribution fits, curves mean it doesn’t.
This flexibility makes probability plots one of the more versatile tools in exploratory data analysis. Rather than committing to a single assumed distribution, you can generate several probability plots for the same dataset, each testing a different candidate. The one that produces the straightest line is typically the best fit for your data.

