How to Read a Study (Without a Statistics Degree)

Most scientific papers follow the same structure, and once you know what each section is actually telling you, you can pull useful information from a study in minutes. The key is knowing where to look first, what the numbers really mean, and which details separate a strong study from a weak one.

Start With the Abstract, Not the Beginning

Almost every modern research paper follows a format called IMRaD: Introduction, Methods, Results, and Discussion. This structure became standard over the twentieth century, and knowing it gives you a roadmap. But you don’t need to read these sections in order.

A practical approach, sometimes called the three-pass method, saves you from spending an hour on a paper that isn’t relevant or reliable. On your first pass, spend five to ten minutes reading only the title, abstract, introduction, section headings, and conclusion. Skip the dense middle entirely. Glance at the references to see if you recognize any foundational work. This tells you whether the paper is worth your time at all.

If it is, your second pass (up to an hour) focuses on figures, graphs, and key data tables. These are where the actual findings live, stripped of the authors’ interpretation. Mark any unfamiliar references for later. A third pass, which most non-specialists never need, involves challenging every assumption and mentally reconstructing the study’s logic from scratch.

What Each Section Actually Tells You

The Introduction frames the question the researchers set out to answer and explains why it matters. Read it to understand context, but know that authors are building a case for their own work here. They’ll emphasize gaps in existing research that conveniently lead to their study.

The Methods section describes exactly what the researchers did: who they studied, how many people were involved, what measurements they took, and how they analyzed the data. This is the section most people skip and the one that matters most. A flashy result built on a flawed method is worthless. Look for how participants were selected, whether there was a comparison group, and whether the study was designed before or after the data was collected.

The Results section presents the raw findings, ideally with specific numbers, tables, and figures. Read this before you read the Discussion, so you can form your own impression of what the data shows.

The Discussion is where the authors interpret their results, compare them to other research, and acknowledge limitations. Pay close attention to the limitations paragraph. Honest researchers will tell you exactly what could be wrong with their own study.

Not All Studies Are Created Equal

Research quality exists on a hierarchy, and knowing where a study falls on that hierarchy tells you how much weight to give it. At the top sit systematic reviews and meta-analyses, which pool data from many individual studies to find patterns across all of them. Below those are randomized controlled trials (RCTs), where participants are randomly assigned to a treatment or comparison group. Next come cohort studies (following groups over time) and case-control studies (comparing people with a condition to people without one). Near the bottom are case reports, which describe what happened to a single patient or a small handful. At the very base: expert opinion and anecdotal evidence.

A single case report describing one patient’s dramatic recovery is interesting, but it tells you almost nothing about whether a treatment works in general. An RCT with hundreds of participants tells you far more. A meta-analysis combining dozens of RCTs tells you the most. When you see a health claim online, check what level of evidence supports it.

How to Read the Numbers Without a Statistics Degree

P-Values

You’ll see “p < 0.05” in nearly every study, and it’s worth understanding what this actually means. A p-value answers a specific question: if there were truly no difference between two groups, how likely is it that we’d see a difference this large just by chance? A p-value of 0.05 means there’s a 5% probability (1 in 20) the result is a fluke. That 0.05 cutoff has been the convention since the early 1900s, when the statistician Ronald Fisher selected it as a convenient decision-making threshold. It stuck, but it’s arbitrary. A p-value of 0.049 isn’t meaningfully different from 0.051, even though one crosses the line and the other doesn’t.

Crucially, a p-value does not tell you the probability that the finding is true. It doesn’t tell you how large or important the effect is, either. A tiny, meaningless difference can be “statistically significant” if the study is large enough.

Confidence Intervals

A 95% confidence interval (CI) gives you a range of values that likely contains the true answer. If a study reports that a treatment reduced blood pressure by 5 points with a 95% CI of 2 to 8, that means the researchers are reasonably confident the true reduction falls somewhere between 2 and 8 points. A narrow interval means the estimate is precise. A wide one means there’s a lot of uncertainty. If a confidence interval crosses zero (say, -1 to 11), the effect might not exist at all.

Absolute Risk vs. Relative Risk

This is where headlines most often mislead. Relative risk compares two numbers as a ratio. Absolute risk tells you the actual difference. Consider this example: a treatment reduces the risk of dying from a disease by 50%. That sounds enormous. But if the original risk was 1 in a million, the new risk is 1 in 2 million. The relative risk dropped by half, but the absolute risk changed by 0.00005%. One fewer person out of two million will be affected. That 50% figure, while technically accurate, is functionally meaningless for any individual person.

Whenever you see a percentage change in a headline, look for the absolute numbers. A “30% increased risk” of something that affects 3 people per 100,000 means it now affects about 4 people per 100,000. That context completely changes how you should feel about it.

Check Who Paid for It

Every reputable journal requires authors to disclose their funding sources and financial conflicts of interest. You’ll typically find this information at the end of the paper, in a section labeled “Disclosures,” “Conflicts of Interest,” or “Funding.” The NIH requires that institutions receiving federal grants maintain and publicly post policies on financial conflicts, and investigators must disclose any significant financial interest that could affect the design, conduct, or reporting of their research.

Industry funding doesn’t automatically invalidate a study, but it’s a factor worth noting. A study on the health benefits of a supplement funded entirely by the supplement manufacturer deserves more scrutiny than one funded by an independent government agency. Look at whether the researchers have financial ties to companies that stand to profit from a positive result.

Red Flags That Signal Weak Evidence

Some problems are easy to spot once you know what to look for.

  • Tiny sample size: Studies with very few participants are more likely to produce unreliable results. Using too small a sample can lead to findings that look significant but don’t hold up when repeated with more people. Inadequate sample size also creates ethical problems, since participants are exposed to an experiment that may not be capable of producing a meaningful answer.
  • P-hacking: This is the practice of running many different statistical tests on the same data and only reporting the ones that produce significant results. Because the probability of a false positive increases with every additional test, a researcher running dozens of comparisons will eventually find something “significant” by sheer chance. Signs include results that seem oddly specific (a drug works only in left-handed women aged 35 to 42) or a paper that tests many outcomes but highlights just one.
  • HARKing: Short for “hypothesizing after the results are known.” This is when researchers look at their data first, find an interesting pattern, and then write their paper as if they predicted that pattern all along. It turns exploratory fishing into what appears to be a confirmed hypothesis.
  • No comparison group: If a study reports that people improved after a treatment but doesn’t compare them to a group that received no treatment (or a placebo), you can’t know whether the treatment caused the improvement. People often get better on their own.
  • Preprint, not peer-reviewed: Preprints are papers posted online before they’ve been reviewed by other scientists. They can contain valuable early findings, but they haven’t been vetted for errors in methodology or analysis. During the COVID-19 pandemic, research showed that the line between preprints and peer-reviewed papers was sometimes blurry, with nearly 20% of papers in one analysis posted as “preprints” only after they had already been accepted by a journal, muddying comparisons between the two.

Preregistration Is a Good Sign

Some researchers preregister their studies, meaning they publicly record their hypothesis, methods, and analysis plan before collecting any data. This makes p-hacking and HARKing much harder, because anyone can compare the finished paper to the original plan. If a study mentions preregistration, that’s a point in its favor.

That said, preregistrations aren’t foolproof. Analysis plans are sometimes described vaguely, leaving room for researchers to quietly adjust their approach without detection. Details like how outliers will be handled or how missing data will be treated are often left unspecified, which opens the door to the same flexibility preregistration was meant to prevent. A more rigorous format called a registered report, where the journal reviews and accepts the study plan before data is collected, offers stronger protection.

Putting It All Together

When you encounter a study, whether in a news article or shared on social media, run through a quick mental checklist. What type of study is it, and where does it fall on the evidence hierarchy? How many people were involved? Was there a control group? Who funded it? Are the reported numbers absolute or relative? Does the confidence interval suggest a precise finding or a vague one? Was the study peer-reviewed?

No single study proves anything. Science works through accumulation. The most trustworthy conclusions come from multiple well-designed studies, ideally synthesized in a systematic review, all pointing in the same direction. A single paper with a surprising finding is a starting point for further investigation, not a final answer.