A one-tailed test is a statistical test that looks for an effect in only one direction. Instead of asking “is there a difference?” it asks “is the result specifically greater than (or specifically less than) a certain value?” This matters because it changes how you set up your hypothesis, how you calculate your p-value, and how much statistical power your test has.
How a One-Tailed Test Works
In hypothesis testing, you start with a null hypothesis (there’s no effect) and an alternative hypothesis (there is an effect). A one-tailed test makes the alternative hypothesis directional. For example, if a light bulb company claims their bulbs last more than 850 hours, the hypotheses would be: the null says the average lifespan equals 850 hours, and the alternative says the average lifespan is greater than 850 hours. The key word is “greater than.” You’re only looking in one direction.
This directional focus changes where you look for evidence against the null hypothesis. With a standard significance level of 0.05, a one-tailed test puts all 5% of that threshold into one end of the distribution. If you’re testing whether something is greater, you’re looking at the top 5%. If you’re testing whether something is less, you’re looking at the bottom 5%. The result is statistically significant only if it falls in that one region.
One-Tailed vs. Two-Tailed Tests
A two-tailed test splits the significance level in half. At a 0.05 threshold, it places 2.5% in each tail of the distribution, checking for effects in both directions. A one-tailed test concentrates the full 5% in a single tail. This difference has real consequences for your results.
The critical values are lower for a one-tailed test. For a standard normal distribution at the 0.05 significance level, you need a z-score of 1.645 to reach significance with a one-tailed test, compared to 1.96 for a two-tailed test. At the stricter 0.01 level, the one-tailed critical value is 2.33 versus 2.576 for two-tailed. This means a one-tailed test can detect a real effect with a smaller sample or a weaker signal, as long as that effect goes in the predicted direction.
The p-values differ too. A one-tailed p-value is exactly half the two-tailed p-value when the result goes in your predicted direction. If a two-tailed test gives you a p-value of 0.008, the one-tailed p-value would be 0.004. But if the effect goes the opposite way from your prediction, the one-tailed p-value flips to 0.996, making it essentially impossible to call that result significant. You’ve committed in advance to ignoring effects in the other direction.
The Power Advantage
One-tailed tests are more powerful than two-tailed tests when the researcher correctly identifies the direction of the effect. “More powerful” means you have a higher probability of detecting a real effect when one exists. Because the critical value is lower (1.645 instead of 1.96 for z-tests at 0.05), results that would fall short of significance in a two-tailed test can cross the threshold in a one-tailed test. This can be especially valuable in studies with small samples where every bit of statistical power counts.
The catch is that this power boost only works if the effect actually goes the direction you predicted. If the true effect is in the opposite direction, a one-tailed test has zero power to detect it. You’ve essentially made a bet, and if you’re wrong, you miss the finding entirely.
When a One-Tailed Test Is Appropriate
Not every situation calls for a one-tailed test, and choosing one after seeing your data is a form of cheating. A classic framework from the psychology literature lays out three criteria. First, a difference in the unpredicted direction would be meaningless for your purposes. Second, results in the unpredicted direction would not change any decision or action you’d take compared to finding no difference at all. Third, your directional prediction comes from established theory, and the opposite result isn’t supported by any competing theory.
The common thread is that you genuinely have no use for a result in the other direction. If a surprise finding in the opposite direction would be interesting or important, you should use a two-tailed test.
One-Tailed Tests in Clinical Trials
One of the clearest real-world applications is in non-inferiority trials. These are clinical studies designed to show that a new treatment is “not meaningfully worse” than an existing one. The null hypothesis says the new treatment is inferior by more than some predefined acceptable margin. The alternative says it’s not that much worse. Researchers only care about one direction: is the new treatment too far behind the standard? They’re not testing whether it’s superior.
Because the hypothesis is one-sided, the data are analyzed with a one-tailed test. Researchers calculate a one-sided 95% confidence interval to find the worst-case difference between treatments. If the lower limit of that interval stays within the acceptable margin, the new treatment is considered non-inferior. The U.S. FDA actually requires a stricter threshold for these trials, mandating a 97.5% confidence interval cutoff, which provides extra protection against false conclusions.
Risks of Misuse
The biggest risk with one-tailed tests is using them to make a borderline result look significant. If a two-tailed test gives you a p-value of 0.07 (not significant at 0.05), switching to a one-tailed test cuts that to 0.035 (now significant). This kind of after-the-fact decision falls under the umbrella of p-hacking, where researchers try different analytical choices until they get a significant result.
P-hacking is a well-documented problem in science. It occurs when researchers selectively report the analyses that produce significant findings while burying the ones that don’t. Choosing a one-tailed test after collecting data is one of many ways this can happen. The arbitrary 0.05 cutoff already incentivizes questionable practices, and the flexibility to halve your p-value by switching test types makes the problem worse.
The solution, supported by research published in 2023, is preregistration. When researchers commit to a one-tailed test before collecting data and publicly register that decision, one-tailed tests control false positives at the same rate as two-tailed tests. The issue isn’t the test itself. It’s the freedom to choose it after seeing the results. If you specify your directional hypothesis in advance, a one-tailed test is a legitimate and more powerful tool.
Choosing Between One-Tailed and Two-Tailed
Use a one-tailed test when you have a strong, theory-based reason to predict a specific direction and you genuinely don’t care about effects in the other direction. Use a two-tailed test when an effect in either direction would be meaningful, when you’re doing exploratory research, or when you’re not sure which direction to expect. When in doubt, a two-tailed test is the safer and more conservative choice, since it won’t miss a surprising result that goes the “wrong” way.

