The curse of dimensionality describes the fact that as you add more dimensions (features, variables, columns) to a dataset, many algorithms become dramatically less effective or impossibly slow. The term was coined by mathematician Richard Bellman in the 1960s to describe how computational complexity of numerical methods increases exponentially with the number of dimensions. While the phrase originated in applied mathematics, it now shows up constantly in machine learning, statistics, and data science, where it creates real problems for anyone working with data that has dozens, thousands, or millions of features.
Why More Dimensions Break Things
The core issue is that high-dimensional space behaves in deeply unintuitive ways. Our brains are wired for three dimensions, and the patterns we rely on in 2D or 3D simply stop working as you scale up. Three specific problems emerge: space becomes impossibly vast, distances lose meaning, and data points retreat into corners.
To understand how fast the space grows, imagine dividing each dimension into 10 bins. In two dimensions, you have a 10×10 grid with 100 cells. In three dimensions, 1,000 cells. By 20 dimensions, you have 10²⁰ cells, more than the estimated number of grains of sand on Earth. Your dataset, no matter how large, becomes hopelessly sparse. Every data point is essentially alone, surrounded by empty space in every direction. This sparsity makes it nearly impossible for algorithms to find reliable patterns because there just aren’t enough nearby examples to learn from.
Distances Stop Working
Many algorithms, particularly classification and clustering methods, rely on measuring how “close” two data points are. In high dimensions, this concept breaks down. The gap between the nearest and farthest points in a dataset shrinks relative to the overall distances. Mathematically, the ratio of the difference between maximum and minimum distances to the minimum distance converges toward zero as dimensions increase. In practical terms, every point looks roughly the same distance from every other point.
This phenomenon, called distance concentration, is devastating for algorithms like k-nearest neighbors. If all distances are nearly equal, asking “which points are closest?” becomes a meaningless question. The algorithm can’t reliably distinguish true neighbors from distant strangers. Research on high-dimensional nearest-neighbor search has also revealed a secondary problem: certain points become “hubs” that appear as nearest neighbors of many other points regardless of whether they’re actually similar. These hubs pollute classification results because they inject incorrect information into predictions for large numbers of other data points.
Volume Hides in the Corners
Here’s one of the stranger facts about high-dimensional geometry: almost all the volume of a cube concentrates in its corners. Picture a square with a circle inscribed inside it. In 2D, the circle takes up about 79% of the square’s area. Now scale that up. Inscribe a sphere inside a cube in higher and higher dimensions. The fraction of the cube’s volume occupied by the sphere plummets toward zero.
A probabilistic proof from UC Davis illustrates this cleanly. If you pick a random point inside a high-dimensional cube, the probability that it also falls inside the inscribed sphere is at most e^(-d/10), where d is the number of dimensions. By 100 dimensions, that probability is essentially zero. Almost every random point in the cube sits out in one of its corners, far from the center. This matters because many algorithms assume data is spread somewhat evenly through a space. In reality, high-dimensional data clusters in corners and along edges, making standard sampling and optimization strategies unreliable.
Computational Costs Explode
Bellman’s original concern was computational. Many optimization and search problems require examining combinations of possibilities across dimensions. Adding a single dimension can multiply the work required. The classic example is the traveling salesman problem: a brute-force solution has a time complexity that grows factorially with the number of cities, and even Bellman’s own dynamic programming improvement only brought it down to roughly 2ⁿ × n², still exponential. This distinction between polynomial growth (manageable) and exponential growth (intractable) is the fundamental dividing line in computational complexity theory, and the curse of dimensionality pushes problems squarely into the exponential side.
For machine learning, this means training time can skyrocket as you add features. But the computational cost is often the lesser problem compared to the statistical one: you need exponentially more training data to maintain the same level of model accuracy as dimensions increase. If you don’t have that data, your model overfits, memorizing noise instead of learning real patterns.
Where It Hits Hardest
Genomics is one of the fields most affected. A typical gene expression study might measure 20,000 genes across only a few hundred patients. This “wide” data, where features vastly outnumber samples, is a textbook curse-of-dimensionality scenario. Running a standard statistical test on each gene individually creates a massive multiple-testing burden that erodes statistical power. Research published in Nature Genetics describes how even popular workarounds like running PCA on high-dimensional clinical data can miss important signals, because PCA assumes linear relationships between features and the underlying biology.
Image recognition faces similar challenges. A modest 100×100 pixel grayscale image lives in 10,000-dimensional space. Color images triple that. Without dimensionality reduction, training a classifier on raw pixel values would require absurd amounts of data. Natural language processing deals with vocabularies of hundreds of thousands of words, each potentially a dimension. Recommendation systems, financial modeling, sensor networks: any domain where the number of measured variables is large relative to the number of observations will encounter these problems.
The Manifold Hypothesis
There’s a reason machine learning works at all despite these challenges. The manifold hypothesis, a widely accepted idea in the field, proposes that real-world high-dimensional data isn’t actually spread across all those dimensions. Instead, it concentrates near a much lower-dimensional surface (a manifold) embedded within the larger space. A photo of a face might exist in a space of millions of pixels, but the set of all plausible face images occupies a tiny, structured region of that space. The “true” dimensionality of face images might be just a few dozen variables capturing pose, lighting, expression, and identity.
This is why dimensionality reduction works. If the data really does live on or near a low-dimensional manifold, you can find that manifold and project the data down to it without losing much information.
Reducing Dimensions in Practice
Two broad strategies exist for fighting the curse: feature selection and feature extraction. They solve different problems and suit different situations.
Feature selection keeps a subset of your original variables and discards the rest. You’re removing redundant, noisy, or uninformative columns. The advantage is interpretability: if you started with 500 features and selected 30, you know exactly what those 30 represent. Filter methods are fast and scale well to large datasets but may miss interactions between features. Wrapper methods test subsets against a specific model and find better combinations, but they’re computationally expensive and can overfit. Embedded methods, which build feature selection into the model training process, tend to strike the best balance. In a comparative study on financial data, embedded techniques produced 83% of the top-performing models.
Feature extraction creates new variables by transforming the originals. PCA, the most common linear method, finds directions of maximum variance in your data and projects everything onto those directions. It’s fast and interpretable, but it assumes linear relationships and struggles with outliers. For nonlinear structure, t-SNE preserves local similarities and is widely used for visualizing single-cell genomics data, though it can distort global structure and is sensitive to parameter choices. UMAP balances local and global faithfulness with better speed. Autoencoders, a deep learning approach, can learn complex nonlinear transformations but need large datasets and sacrifice transparency.
A common hybrid approach runs PCA first to denoise and decorrelate the data, then applies UMAP or t-SNE for nonlinear refinement. This combination improves scalability and produces more stable results on large datasets. When choosing how many dimensions to keep, a standard rule of thumb is retaining enough PCA components to explain 90-95% of the total variance. More sophisticated methods estimate the intrinsic dimensionality of the data, essentially asking: how many dimensions does this manifold actually have?
When You Should Worry About It
The curse of dimensionality isn’t equally dangerous in every situation. It matters most when your number of features is large relative to your number of samples. A dataset with 50 features and 1,000,000 rows is rarely cursed. A dataset with 10,000 features and 200 rows almost certainly is. Distance-based methods like k-nearest neighbors and k-means clustering are the most vulnerable. Tree-based models like random forests and gradient-boosted trees handle high dimensions better because they implicitly perform feature selection at each split. Deep neural networks, when given enough data, can learn useful representations that sidestep the curse, which is essentially what the manifold hypothesis predicts.
If your model’s performance on training data is great but its performance on new data is poor, high dimensionality may be the culprit. That gap between training and test performance is overfitting, and it’s one of the most common symptoms. Reducing dimensions, either by selecting features or extracting new ones, is often the single most effective step you can take before tuning anything else.

