P-hacking is the practice of manipulating data analysis until a statistically significant result appears, even when no real effect exists. It exploits the flexibility researchers have in how they collect, analyze, and report data, turning what should be an objective process into something closer to a treasure hunt for a specific number: a p-value below 0.05.
To understand why this matters, you need to know what that number means. A p-value is supposed to tell researchers how likely their results would be if nothing interesting were actually going on. The lower the p-value, the more surprising the result. By convention, anything below 0.05 (a 1-in-20 chance) gets labeled “statistically significant.” That threshold traces back to the statistician Ronald Fisher, who in the 1950s called it a “convenient” cutoff. He meant it as a rough guideline, not a rigid gate. But it became one, and that gate now determines what gets published, funded, and reported as fact.
Why the 0.05 Threshold Creates a Problem
When you run a single statistical test with nothing real going on, there’s a 5% chance you’ll get a false positive: a result that looks significant but isn’t. That sounds manageable. But the math changes fast when you run multiple tests. With 10 comparisons, the probability of at least one false positive jumps to 40%. At 13 comparisons, it hits 50%. This is known as the multiple comparisons problem, and p-hacking exploits it directly.
Researchers rarely test just one thing. They might measure dozens of outcomes, compare multiple groups, or slice their data in various ways. Each of those choices is another roll of the dice. If you only report the rolls that came up in your favor, you can make almost anything look statistically significant.
How Researchers P-Hack
P-hacking isn’t one technique. It’s a collection of practices, some deliberate and some that researchers fall into without realizing it.
Data peeking. A researcher checks results repeatedly as data comes in and stops collecting once the numbers cross the 0.05 line. If they’d kept going, the effect might have vanished. This is sometimes called “optional stopping,” and it’s one of the most common forms because it feels harmless: why keep spending money on a study when you already have your answer? The problem is that small samples are noisy, and stopping at a lucky moment locks in that noise as a “finding.”
Selective reporting. A study might measure ten different outcomes, but only the two that reached significance make it into the paper. The eight null results disappear. From the outside, it looks like the researchers tested a specific hypothesis and confirmed it. In reality, they cast a wide net and kept only what they caught. This also happens at the study level: researchers who run multiple experiments sometimes shelve the ones that didn’t work and publish only the ones that did.
Tweaking the analysis. Researchers have enormous flexibility in how they set up their statistical models. One powerful lever is choosing which variables to “control for.” By adding or removing these background factors, a borderline result can tip over into significance. Comparisons of doctoral dissertations with their later published versions have shown this in action: authors added new conditions to their analyses between the dissertation and the journal article, and those additions conveniently produced more significant results.
Subgroup fishing. If the overall result isn’t significant, a researcher can start slicing the data by age, sex, income, location, or any other variable until a subgroup shows the desired effect. This might be reported as the key finding rather than the overall null result.
A Famous Demonstration
In 2011, a team of researchers led by Joseph Simmons published a paper with a deliberately absurd finding: listening to the Beatles song “When I’m Sixty-Four” made people younger. Not “feel younger.” Actually younger, by measurable calendar age. They achieved this through a combination of flexible data collection and selective reporting, all using standard practices that journals routinely accepted at the time. Their point was stark: “it is unacceptably easy to accumulate and report statistically significant evidence for a false hypothesis.” Their simulations showed that in many cases, a researcher is more likely to falsely find evidence for a nonexistent effect than to correctly find no effect at all.
The Replication Crisis
P-hacking is one of several forces behind what scientists call the replication crisis: the discovery that a troubling number of published findings can’t be reproduced by independent teams. In social psychology, less than 30% of results held up when replicated. Cognitive psychology fared better but still lost about half. When researchers at the biotech firm Amgen tried to confirm 53 landmark studies in preclinical cancer research, only six held up.
The relationship between p-hacking and failed replications is real but complicated. Some analyses suggest that a low rate of true effects in certain fields, not p-hacking alone, is the main driver of poor replication. In other words, the problem isn’t just that researchers are gaming their analyses. It’s also that many fields are testing hypotheses that were unlikely to be true in the first place, and p-hacking makes it easier for those weak hypotheses to pass the significance filter.
What a P-Value Actually Tells You
Part of the problem is widespread misunderstanding of what p-values mean. In 2016, the American Statistical Association took the unusual step of issuing a formal statement with six principles about p-values. The core messages: a p-value does not measure the probability that your hypothesis is true. It does not measure the size or importance of an effect. Scientific conclusions should not rest solely on whether a p-value crosses 0.05. And a p-value by itself does not provide good evidence for or against a hypothesis.
These points might seem obvious when stated plainly, but they cut against decades of practice. Entire careers, funding decisions, and medical guidelines have been built on the binary question of whether p was above or below 0.05.
How the Problem Is Being Addressed
The most widely adopted countermeasure is preregistration: researchers publicly record their hypotheses, methods, and analysis plan before they collect data. This makes it much harder to quietly shift the goalposts after seeing results. A stronger version, called registered reports, goes further. Journals review and accept (or reject) a study based on its design before any data exists. The results, whatever they turn out to be, get published regardless of whether they’re significant. This removes the incentive to hack results entirely, because publication no longer depends on hitting 0.05.
Statistical methods are also evolving. The most frequent expert recommendations include dropping the phrase “statistically significant” altogether, reporting effect sizes (how big is the difference, not just whether it exists), using confidence intervals to show the range of plausible values, and adopting Bayesian methods that let researchers incorporate prior knowledge into their analysis rather than treating every study as if it exists in a vacuum.
None of these fixes are universal yet. Preregistration is growing but not required by most journals. Many researchers still rely on the 0.05 threshold as their primary decision tool. But the conversation has shifted. Twenty years ago, p-hacking was widespread and largely invisible. Today it has a name, a body of evidence documenting its effects, and a growing set of tools designed to prevent it.

