Collinearity is bad because it makes your regression model unreliable for understanding which variables actually matter. When two or more predictor variables are highly correlated with each other, the model can’t cleanly separate their individual effects. The overall predictions may still be fine, but the coefficients for each variable become unstable, sometimes wildly so.
This matters most when you’re trying to answer questions like “does this variable have an effect?” If you only care about generating predictions, collinearity is less of a problem. But if you’re interpreting coefficients, building explanations, or making decisions based on which variables are significant, collinearity can lead you to completely wrong conclusions.
What Collinearity Actually Does to Your Model
In a regression model, each coefficient represents the effect of one variable while holding the others constant. When two predictors are highly correlated, “holding one constant while changing the other” becomes nearly impossible in practice. The data simply doesn’t contain enough independent variation to tease apart their separate contributions.
The result is that the model has many nearly equivalent ways to split the credit between the correlated variables. Small changes in the data, like adding a few observations or removing a few, can cause the coefficients to swing dramatically. One variable might absorb most of the effect in one sample and very little in another. Coefficients can even flip signs, showing a positive relationship when the true effect is negative, purely because the correlated partner is soaking up more than its share.
Critically, this instability doesn’t show up in your model’s overall fit. The R-squared stays the same, and the model’s predictions remain accurate. As one epidemiology review put it, multicollinearity does not affect the overall fit or the predictions of the model. The damage is entirely to the individual coefficients and their interpretation.
Standard Errors Inflate, Significance Disappears
The most concrete consequence of collinearity is inflated standard errors. When the model can’t precisely estimate a coefficient, the uncertainty around that estimate grows. Larger standard errors mean wider confidence intervals, which means your p-values get larger. A variable that genuinely predicts the outcome can appear statistically insignificant simply because it’s correlated with another predictor.
This is how collinearity leads to misleading statistical results. You might conclude that neither variable matters when in fact both do. Or you might conclude that one matters and the other doesn’t, when the “winning” variable just happened to capture slightly more of the shared variance in your particular sample. The problem isn’t that your model is wrong in a detectable way. It’s that it gives you false confidence about which specific variables are driving the results.
When Collinearity Matters (and When It Doesn’t)
Whether collinearity is actually a problem depends entirely on what you’re using the model for. If your goal is prediction, you typically don’t need to worry. As long as the relationships between predictors remain similar in new data, the model will predict just as well even with correlated inputs. Two redundant variables aren’t a problem if you don’t care which one gets the credit.
If your goal is inference, collinearity is a serious threat. Inference means you want to understand the independent contribution of each variable: does body fat percentage affect bone density after accounting for weight? Does education level predict income independently of occupation? When predictors share overlapping information, the model can’t cleanly answer these questions. The common interpretation of a regression coefficient, measuring the expected change when you increase one variable by one unit while holding everything else constant, becomes practically impossible when the “everything else” can’t realistically be held constant because it moves in lockstep with the variable you’re changing.
Real-World Variables That Cause Problems
Collinearity shows up constantly in real data because many things we measure are naturally related. Height and weight. Education and income. Whole grain consumption and refined grain consumption (people who eat more of one tend to eat less of the other). In economics, labor and capital inputs are often so tightly correlated that separating their effects is nearly impossible: one analysis found a correlation of 0.985 between them, producing variance inflation factors above 35.
There’s also a type of collinearity you create yourself. Adding an interaction term (like weight multiplied by body fat percentage) or a squared term (like age and age-squared) to your model introduces correlation by construction. This is called structural multicollinearity, and it’s generally the less troublesome kind because you can often fix it by centering your variables before creating the interaction or polynomial terms. Data-based multicollinearity, the kind that comes from the natural structure of your data, is harder to resolve because it reflects the actual world your data came from.
How to Detect It
The most common diagnostic is the variance inflation factor, or VIF. It measures how much the variance of a coefficient is inflated due to correlation with other predictors. A VIF of 1 means no collinearity at all. VIFs above 4 warrant a closer look, and VIFs above 10 indicate serious collinearity that likely needs to be addressed. Some researchers use a stricter threshold of 5 as the cutoff for concern.
Another diagnostic is the condition index, which examines the overall structure of the predictor matrix. Values above 10 suggest moderate collinearity, and values above 30 indicate severe collinearity. Tolerance, which is simply the inverse of VIF, flags problems when it drops below 0.1 to 0.2.
Simply checking the correlation matrix between your predictors is a reasonable first step but can miss multicollinearity that involves three or more variables acting together. Two variables might each have only moderate pairwise correlations, yet jointly create collinearity problems. VIF catches this; a correlation table doesn’t.
What You Can Do About It
The right fix depends on why the collinearity exists and what your model is for. If you created it through interaction or polynomial terms, centering your variables (subtracting the mean) before computing those terms often resolves the issue. If two predictors genuinely measure nearly the same thing, dropping one of them is the simplest solution. You lose some nuance but gain interpretable coefficients.
Combining correlated variables into a single index or composite score is another option. Principal component analysis does this formally, creating new uncorrelated variables from combinations of the original ones. The tradeoff is that the new variables are harder to interpret in plain terms.
Collecting more data helps in some situations, particularly when collinearity is a byproduct of a small sample. With more observations, you’re more likely to have cases where the correlated variables diverge enough for the model to separate their effects. But if the variables are fundamentally linked in the population, no amount of data will solve the problem.
If your goal really is just prediction, the most practical response is often to simply acknowledge the collinearity, avoid interpreting individual coefficients, and move on. Regularization techniques like ridge regression can also stabilize coefficients by deliberately introducing a small amount of bias, trading perfect coefficient estimates for much better stability.

