Tracking sensitivity refers to how well a system detects what it’s designed to detect. The term appears across several fields, from wearable fitness devices to medical screening to visual perception research, but the core idea is the same: it measures the proportion of real events a tracking system correctly identifies. A sleep tracker with a sensitivity of 0.95, for example, correctly identifies 95% of the moments you’re actually asleep.
The Basic Formula
Sensitivity is calculated as true positives divided by the sum of true positives and false negatives. In plain terms, that means: out of everything that actually happened, how much did the system catch? If a heart rate monitor detects 90 out of 100 actual heartbeats, its sensitivity is 0.90 (or 90%). The events it missed are false negatives.
Sensitivity has an inherent tradeoff with specificity, which measures how well a system avoids false alarms. As sensitivity increases, specificity tends to decrease, and vice versa. A motion sensor cranked up to catch every possible movement will also flag things that aren’t real movement. A more conservative sensor will produce fewer false alarms but miss more genuine events. Designers of tracking systems are constantly balancing these two qualities depending on the purpose of the device or test.
Tracking Sensitivity in Wearable Devices
For fitness trackers and smartwatches, tracking sensitivity describes how reliably the device picks up the biological signals it monitors, whether that’s steps, heart rate, or sleep stages. Several factors influence this.
Sensor quality is the most obvious one. Wrist-based heart rate monitors use light sensors that read blood flow through your skin. The quality of the signal depends on whether the sensor can clearly identify the peak of each heartbeat pulse. Researchers categorize these signals into two tiers: basic-quality pulses, where the main heartbeat peak is visible and heart rate can be estimated, and high-quality pulses, where the full waveform is clean enough for deeper analysis like heart rate variability.
How you wear the device matters just as much as the hardware inside it. Loose skin contact, incorrect placement on the wrist, and motion during exercise all introduce noise that degrades the signal. Sensor degradation over time, Bluetooth synchronization errors, and gaps from simply not wearing the device also contribute to missing or inaccurate data. A rapid systematic review in JMIR mHealth uHealth identified user error (misplacement, poor skin contact) and hardware issues (sensor malfunction, degradation) as the leading causes of poor data quality from wearables.
For accelerometers, which track movement in fitness bands and research-grade activity monitors, sensitivity thresholds are measured in milli-G (mg) units. Researchers have found that acceleration below about 13 mg with a range under 50 mg on at least two axes reliably indicates the device is sitting still, not being worn. These tiny thresholds show just how sensitive modern accelerometers are, and why even minor vibrations can be misread as activity.
How Sensitive Are Consumer Sleep Trackers?
Sleep tracking is one area where sensitivity has been directly measured against a clinical gold standard. Polysomnography (PSG) is the lab-based sleep study that monitors brain waves, eye movement, and muscle activity overnight. When seven popular consumer devices were compared against PSG, all achieved sensitivity scores of 0.93 or higher for detecting sleep versus wakefulness.
The Garmin Fenix 5S and Garmin Vivosmart 3 scored highest at 0.99, meaning they correctly identified 99% of the 30-second intervals where the person was actually asleep. The Fitbit Alta HR hit 0.95, the Actiwatch reached 0.97, and the ResMed S+ came in at 0.93. These numbers sound impressive, but there’s a catch: high sensitivity for detecting sleep often comes with lower specificity for detecting wakefulness. In other words, these devices are great at knowing when you’re asleep but less reliable at knowing when you’re lying awake in bed.
Tracking Sensitivity in Public Health
In disease surveillance, tracking sensitivity measures how well a monitoring system detects cases of a health condition as they occur in a population. The CDC defines it as the ability of surveillance to detect the health problem it is intended to detect. A flu tracking system with high sensitivity catches most flu cases in a community. One with low sensitivity misses a large share of them, potentially delaying public health responses.
Most surveillance systems detect only a fraction of cases that actually occur. The practical question isn’t whether a system catches every single case, but whether it’s sensitive enough to be useful for prevention and control. A system that picks up 40% of cases might still reveal outbreak patterns and seasonal trends effectively, even though it misses the majority of individual cases.
Tracking Sensitivity in Vision Research
Cognitive scientists use the term differently. In visual perception research, tracking sensitivity refers to how precisely a person can follow a moving target with their eyes or a cursor. Researchers at the Journal of Vision developed a method where participants track a randomly moving blob on screen using a mouse. The gap between where the target actually is and where the person places their cursor reveals their “positional uncertainty,” a measure of how sensitive their visual system is to the target’s location.
This approach uses a statistical model (called a Kalman filter) to separate genuine visual limitations from motor control noise. The result is an estimate of visual sensitivity that’s comparable to traditional lab tests but can be collected continuously rather than through repetitive yes-or-no trials. It’s primarily a research tool, used to study how factors like contrast, blur, or attention affect the visual system’s ability to track objects in real time.
Why Tracking Sensitivity Varies
Across all these domains, a few common themes determine sensitivity. The quality of the sensor or test matters, but so does the environment it operates in. A perfectly calibrated heart rate sensor still produces poor data on a loosely worn watch during a sprint. A highly sensitive disease surveillance system still misses cases if clinicians don’t report them.
The lack of standardization across devices and systems is another persistent issue. Different wearable brands place sensors in different positions, use different proprietary algorithms, and define the same measurements in slightly different ways. Two devices can report the same metric, like resting heart rate or sleep duration, while using completely different sensitivity thresholds under the hood. When comparing tracking data across devices or over time after switching brands, these hidden differences in sensitivity can explain why your numbers suddenly shift even though nothing about your body changed.

