The true experimental design is the only research design that allows cause-and-effect conclusions. If you’re looking at a list that includes correlational, descriptive, case study, quasi-experimental, and experimental designs, the answer is the true experiment. What makes it unique is the combination of three features: the researcher actively manipulates a variable, participants are randomly assigned to groups, and a control group provides a baseline for comparison.
Three Requirements for Causal Claims
A cause-and-effect conclusion requires satisfying three conditions that only a true experiment can meet simultaneously. First, the cause must come before the effect (temporal precedence). Experiments handle this by measuring participants on the outcome before introducing any treatment, establishing a baseline, and then measuring again afterward. Survey and observational research often collects all data at one time point, making it impossible to know which variable came first.
Second, there must be a real relationship between the variables, not just a coincidence. Statistical tests built into experimental designs verify this.
Third, and most critically, the cause must be isolated. The researcher needs to rule out the possibility that some outside variable is actually driving the result. This is where random assignment and control groups do the heavy lifting, and where other designs fall short.
Why Random Assignment Matters
Random assignment is the single most important feature separating true experiments from every other design. When participants are randomly placed into treatment or control groups, the process eliminates selection bias and balances the groups on both known and unknown characteristics. Age, personality, health status, motivation, socioeconomic background: all of these potential confounding variables get distributed roughly equally across groups without the researcher needing to identify or measure every one of them.
This matters because if the groups are equivalent before the treatment starts, any difference observed afterward can be attributed to the treatment itself rather than to pre-existing differences between the people in each group. Randomization also ensures that neither the researcher nor the participant knows the group assignment in advance, which prevents conscious or unconscious steering of certain people toward certain conditions.
In studies where specific characteristics like age could heavily influence results, stratified randomization goes a step further. Participants are first sorted by the characteristic (for example, age brackets), then randomly assigned within each bracket, so both groups end up with similar age distributions.
The Role of Control Groups
A control group acts as the counterfactual: it shows what would have happened without the treatment. Without one, there is no way to separate the effect of a treatment from natural changes, the passage of time, or outside influences.
Consider a study testing whether an exercise program reduces stress. If everyone in the study exercises and stress drops, you might assume the exercise worked. But stress levels fluctuate naturally. People’s routines change. Seasons shift. A control group that does not receive the exercise program reveals whether stress also dropped for people who did nothing, which would suggest the change had nothing to do with the exercise. Only when the treatment group outperforms the control group at the end of the study can researchers confidently say the treatment caused the difference.
Why Correlational Designs Cannot Show Causation
Correlational studies measure two or more variables and look for relationships between them, but they fail the causation test for two reasons. They cannot establish which variable came first, and they cannot rule out third variables.
A classic example: research finds that happier people tend to be more generous. Does happiness cause generosity? Does generosity cause happiness? Or does a third factor, like wealth, cause both? A correlational design has no way to distinguish between these possibilities. The researcher observes the relationship but never manipulates anything, so the direction of the effect and the presence of hidden variables remain unknown. This is the core of the “correlation does not mean causation” principle.
Where Quasi-Experiments Fall
Quasi-experimental designs sit between correlational studies and true experiments. They include some form of treatment or intervention, but they lack random assignment. A researcher might compare students in two different schools, one that adopted a new teaching method and one that didn’t, without being able to randomly assign students to schools.
These designs can suggest causal relationships, and researchers use statistical techniques like regression and propensity-score matching to try to account for group differences. But compared to true experiments, quasi-experiments rely on stronger and more numerous assumptions. Causal conclusions from quasi-experiments are weaker because the groups may differ in systematic ways the researcher cannot fully control for. In some cases, causal effects can only be estimated for specific subgroups rather than the full population.
Threats That Undermine Causal Conclusions
Even well-designed experiments can fail to support causal claims if certain threats to internal validity are not controlled. Knowing these threats helps you evaluate whether a study’s conclusions are trustworthy.
- Confounding (selection bias): Pre-existing differences between groups explain the result instead of the treatment. Random assignment directly addresses this.
- Ambiguous temporal precedence: It is unclear whether the supposed cause actually came before the effect. This is common in cross-sectional studies.
- History: An outside event occurs during the study that affects the outcome. For instance, a public health campaign launches midway through a nutrition study.
- Maturation: Natural changes over time (aging, development, healing) are mistaken for treatment effects.
- Regression to the mean: Participants selected because of extreme scores naturally score closer to average on the next measurement, creating the illusion of improvement.
- Attrition: Participants drop out unevenly between groups, skewing the results. If sicker patients leave the treatment group because of side effects, the remaining group looks healthier than it really is.
- Testing effects: The act of measuring the outcome changes behavior. A person weighed at the start of a weight-loss study might eat less simply because they were weighed, regardless of the actual intervention.
True experiments, through randomization, control groups, and careful design, address more of these threats simultaneously than any other research design.
The Evidence Hierarchy
Research designs are often ranked in a pyramid based on how much confidence they provide. At the top sit systematic reviews and meta-analyses, which pool data from multiple high-quality studies (usually randomized controlled trials) to draw the broadest conclusions. Directly below are individual randomized controlled trials. Further down come cohort studies, case-control studies, case series, and expert opinion.
Randomized controlled trials occupy the highest tier for individual studies precisely because they are the only single-study design that establishes causation rather than merely suggesting it. When ethical or practical constraints make a true experiment impossible (you cannot randomly assign people to smoke for 30 years), researchers rely on observational data and apply frameworks like the Bradford Hill criteria, which use factors like strength of association, consistency across studies, dose-response relationships, and biological plausibility to build a circumstantial case for causality. These criteria helped establish the link between smoking and lung cancer decades before any experiment could. But the gold standard, when it can be applied, remains the randomized controlled trial.

