Voluntary response samples are unreliable because they systematically attract people with strong opinions while the majority with moderate or neutral views stay silent. This creates a sample that looks nothing like the actual population it claims to represent. The problem isn’t sample size or poor questions. It’s that the very act of letting people choose whether to participate guarantees the data will be skewed.
The Self-Selection Problem
In a well-designed study, every person in the target population has a known chance of being selected. A voluntary response sample flips this entirely: participants decide for themselves whether to show up. That single difference undermines everything that makes statistical sampling work.
People don’t volunteer at random. They volunteer because something about the topic activates them. Maybe they had a terrible experience. Maybe they’re passionate advocates. Maybe they’re seeking validation or even indirect help. A study published in PLOS One found that volunteers for psychological research consistently showed higher rates of personality and affective disorders compared to non-volunteers. The researchers concluded that the field “may be conducting research on an atypically disordered and motivated group of people, leading to biased views of the reality of psychological effects.” The participants weren’t representative of the general population. They were representative of the kind of person who volunteers for a psychology study.
This is what statisticians call self-selection bias, and it’s baked into the method. You can’t fix it with a bigger sample, because adding more self-selected people just gives you a larger collection of the same skewed group.
Extreme Voices Drown Out the Middle
The most predictable pattern in voluntary response data is that it overrepresents the extremes. People who are thrilled or furious are far more likely to respond than people who feel fine. Think about online product reviews: the one-star and five-star ratings pile up while the three-star “it was okay” experience goes unrecorded. That silent majority with moderate opinions is exactly the group you need to hear from to get an accurate picture, and they’re the least likely to participate.
Consider a company that sends out a post-launch survey. The customers who loved the product respond. The customers who had a terrible experience respond. The large middle group, people who found the product decent but unremarkable, mostly ignores the email. The company then looks at its data and sees a polarized audience, when in reality most customers were somewhere in between. Decisions made from that data will be calibrated for extremes that don’t reflect the typical customer.
Two Famous Examples of How Wrong It Can Go
Advice columnist Ann Landers once asked her readers: “If you had it to do over again, would you have children?” About 10,000 people responded, and 70 percent said no. The result made headlines. But when newspapers and researchers later conducted properly sampled surveys of American parents, 90 to 95 percent said they would have children again. The Landers poll wasn’t just slightly off. It was almost the inverse of reality, because the parents motivated enough to write in were disproportionately the ones with regrets.
A similar problem sank the credibility of Shere Hite’s book “Women and Love.” Hite mailed questionnaires to 100,000 women about their relationships with men. Only 4.5 percent responded. The women who did respond were, by later analysis, disproportionately angry and fed up with men. That anger became the defining theme of the book, presenting a picture of American women’s experiences that was driven almost entirely by who chose to fill out the form.
You Can’t Calculate a Real Margin of Error
One of the less obvious problems with voluntary response samples is that standard statistical tools simply don’t apply to them. The margin of error you see reported in polls (plus or minus 3 percent, for example) relies on a specific mathematical assumption: that every person in the population had a known probability of being selected. When people self-select, that assumption collapses.
As Columbia University’s statistics blog has noted, most researchers understand that applying margin of error to non-probability samples is inappropriate. The issue is that you don’t know the selection mechanism that determined who responded and who didn’t. Without that knowledge, you can’t quantify how far your results might be from the truth. You’re left with numbers that look precise but have no mathematically valid measure of their own accuracy. Any conclusion drawn from the data is really only a conclusion about the subgroup that chose to participate, not about the broader population.
Online Surveys Make the Problem Worse
The internet has made voluntary response sampling easier than ever and, in the process, amplified its flaws. Online surveys get distributed through emails, social media posts, and website pop-ups. There is often no defined population frame at all. You can’t describe who could have seen the survey, which means you can’t describe who the results apply to.
A review in the Indian Journal of Psychological Medicine identified two compounding problems with online surveys specifically. First, they’re limited to people who are literate and have internet access, which immediately excludes significant portions of many populations. Second, only people with sufficient interest or bias in the topic will take the time to click through and respond. The reviewers used the example of a survey about a medical procedure: patients who were traumatized by the procedure are far more likely to respond than patients whose experience was uneventful. The traumatized patients want to be heard. Everyone else has no particular motivation to participate. The result is data that overstates negative outcomes.
Because there’s no way to know the motives of each respondent, there’s also no way to measure or correct for the extent of the bias. The results have to be treated as tentative at best.
Who Gets Left Out Entirely
Beyond attracting the wrong mix of respondents, voluntary response samples also systematically miss entire groups. A website survey only reaches people who visit that website. A social media poll only reaches followers of a particular account. A phone-in poll excludes people without phones or people who don’t pick up calls from unknown numbers.
This is called undercoverage, and it layers on top of the self-selection problem. Even among people who could theoretically access the survey, participation skews toward those with more free time, stronger digital literacy, or greater personal investment in the topic. The result is a sample that’s unrepresentative in two ways at once: it misses whole demographic groups, and among those it does reach, it oversamples the most motivated.
Why Random Sampling Solves This
The fix isn’t complicated in theory, though it’s harder and more expensive in practice. A simple random sample selects people from the full population regardless of whether they would have volunteered. It includes the indifferent, the busy, and the people with no strong opinion. That’s precisely what makes it work. The conclusions you draw from a random sample can be generalized to the whole population because the sample was drawn from the whole population.
Voluntary response samples will always be biased because they only include people who chose to show up. A random sample includes people whether or not they would have chosen to participate, which is what makes the math behind confidence intervals and margins of error valid. The distinction isn’t a technicality. It’s the difference between measuring what a population actually thinks and measuring what its most vocal members are willing to tell you.

