Color photography works by separating light into its red, green, and blue components, recording each one independently, then combining them back into a full-color image. This core principle has remained the same since 1861, whether the recording medium is chemical film or a digital sensor. The methods differ dramatically, but the logic is identical: capture what each primary color is doing in a scene, then layer that information together.
The Idea Behind It All
In 1861, Scottish physicist James Clerk Maxwell demonstrated the concept that makes color photography possible. He asked photographer Thomas Sutton to take three separate black-and-white photographs of a tartan ribbon, each shot through a different colored filter: red, green, and blue. They printed the images on glass and projected all three simultaneously onto a wall. The overlapping projections produced a recognizable color image. Maxwell’s insight was that any visible color can be reconstructed from the right combination of red, green, and blue light, the three primary colors your eyes use to see the world. Every color camera built since, from a 1950s Kodak to your phone, applies this same principle.
How Light Becomes Color
Visible light is a spectrum. When white light hits an object, some wavelengths get absorbed and others bounce back. A red apple absorbs most blue and green wavelengths and reflects red ones. Your eye has three types of color-sensing cells, each tuned roughly to red, green, or blue. Your brain blends the signals from all three into the colors you perceive.
Color photography mimics this biological system. It uses three separate channels, one for red, one for green, one for blue, and records how much of each is present at every point in the scene. The result is three grayscale “maps” of intensity. Merge them and you get a full-color image.
How Color Film Captures an Image
A strip of color film is far more complex than it looks. Inside its thin coating are multiple light-sensitive layers stacked on top of each other, each designed to respond to a different part of the spectrum. From top to bottom, the structure runs like this: the blue-sensitive layer sits on top (because all silver halide, the light-sensitive compound in film, naturally reacts to blue light). Beneath that is a yellow filter layer whose job is to block blue light from reaching the layers below, preventing unwanted contamination. Next comes the green-sensitive layer, and at the bottom sits the red-sensitive layer.
When you press the shutter, light passes through the lens and hits these layers in order. Blue wavelengths are captured at the top, green in the middle, red at the bottom. Each layer records a latent image, an invisible chemical change in its silver halide crystals wherever light struck them.
Turning Chemistry Into Color
The latent images don’t become visible until the film is developed. During this process, a chemical developer reacts with the exposed silver halide grains in each layer. As it does, the developer oxidizes, and the oxidized developer then reacts with dye couplers embedded in each layer. These couplers are color-specific: the blue-sensitive layer produces yellow dye, the green-sensitive layer produces magenta dye, and the red-sensitive layer produces cyan dye. The silver is then bleached away, leaving only the colored dyes behind.
This might seem backward. Why does the blue layer make yellow dye? Because color film uses subtractive color mixing. Instead of adding light (which is what screens do), film works by subtracting wavelengths from white light. Yellow dye absorbs blue, magenta absorbs green, and cyan absorbs red. When light passes through the developed film (or reflects off a print), the combination of these three dye layers filters white light into the correct colors. It’s a system that starts with all the light and selectively removes what doesn’t belong.
How a Digital Sensor Captures Color
A digital camera sensor is a flat grid of millions of tiny light-sensitive sites called photosites, often referred to as pixels. Each photosite can measure how much light hits it, but on its own, it’s colorblind. It records brightness, not color. To add color information, manufacturers place a mosaic of microscopic color filters over the sensor, one filter per photosite.
The most common arrangement is called a Bayer filter. It tiles the sensor with a repeating pattern of red, green, and blue filters. Each group of four pixels contains one red filter, one blue filter, and two green filters. Green gets double the representation because human vision is most sensitive to green light, so weighting the data toward green produces the most natural-looking images. The result is that each pixel in the raw capture records only one color channel. A pixel under a red filter knows how much red light hit that spot, but it knows nothing about the blue or green light there.
Filling In the Missing Colors
Since each pixel only captures one-third of the color information, the camera’s processor has to reconstruct the rest. This process is called demosaicing. It works by examining the values of neighboring pixels and using interpolation algorithms to estimate the missing red, green, and blue values at every single pixel location. A pixel that only recorded red light gets its green and blue values estimated from the green-filtered and blue-filtered pixels surrounding it.
Simple linear interpolation works well enough for video, where speed matters. For high-resolution still images, cameras and editing software use more sophisticated algorithms that analyze gradients and edges in the image to avoid artifacts like false color fringing or staircase patterns along diagonal lines. Some modern approaches use machine learning models trained on millions of images to produce even cleaner results. By the end of this processing step, every pixel has a complete set of red, green, and blue values, and the image appears in full color.
An Alternative Approach: Stacked Sensors
There is one sensor design that skips the Bayer mosaic entirely. Instead of placing a single color filter over each pixel, it stacks three light-sensitive layers on top of each other at every pixel location, much like color film does. Each layer captures a different wavelength range: red, green, or blue. This means every pixel records all three color channels directly, with no estimation required.
The practical advantage is the elimination of demosaicing artifacts and the computational overhead that comes with interpolation. Images from stacked sensors can resolve extremely fine detail without the color fringing that sometimes appears with Bayer sensors. The tradeoff is that stacked sensors are harder and more expensive to manufacture, so Bayer-pattern sensors remain the standard in the vast majority of cameras.
From Sensor Data to a Viewable Image
Once the camera has a full RGB value for every pixel, it needs to turn that raw data into something you can actually look at. This involves several steps that happen almost instantly inside the camera (or later, if you shoot in RAW format and edit on a computer).
First, the camera applies a white balance adjustment, correcting for the color temperature of the light source so that objects that appear white in person also look white in the photo. Fluorescent lights, sunlight, and candlelight all have different color casts, and this step neutralizes them. Next, the camera maps the sensor’s data to a standard color space, typically sRGB for consumer use, which defines exactly how each combination of red, green, and blue values should translate to a visible color on screen.
Bit depth determines how many distinct tones each color channel can represent. At 8 bits per channel, each channel can hold 256 levels of intensity. With three channels, that yields roughly 16.7 million possible color combinations, which is what a standard JPEG delivers. RAW files from cameras typically capture 12 or 14 bits per channel, providing thousands of tones per channel and far more flexibility for adjusting exposure and color in editing before the image is compressed down to 8 bits for display.
Displaying Color: Additive vs. Subtractive
How a color photo reaches your eyes depends on the output medium, and the two systems work in opposite directions. Screens, whether on phones, computers, or televisions, use additive color. Tiny red, green, and blue light elements glow at varying intensities. When all three glow at full brightness, you see white. When all are off, you see black. Mixing red and green light produces yellow. Mixing all three at different levels creates the full range of visible color. Because additive color uses transmitted light, screen colors appear vivid and can represent millions of distinct hues.
Printed photographs use subtractive color. Inks in cyan, magenta, yellow, and black (CMYK) are layered onto white paper. Each ink absorbs certain wavelengths and reflects the rest. The more ink you add, the darker the result, because more light is being absorbed. This is why printed photos generally look slightly more muted than the same image on a bright screen. The physics of reflected light simply can’t match the intensity of light being beamed directly into your eyes.
Both systems, though, trace back to the same starting point: separating light into its component colors and recombining them in a controlled way. Whether that separation happens in layers of dyed gelatin or behind a mosaic of microscopic filters, color photography is fundamentally a trick of decomposition and reconstruction, splitting white light apart and putting it back together to recreate what the world looked like in that fraction of a second.

