De-identified means that personal information has been stripped from a dataset so that the remaining data can no longer be traced back to a specific individual. The term comes up most often in healthcare, where hospitals, insurers, and researchers handle sensitive medical records that are protected by privacy laws. Once health data is properly de-identified, it falls outside most privacy regulations and can be shared, analyzed, or published without individual consent.
The concept sounds simple, but the details matter. What counts as “properly” de-identified, which pieces of information must be removed, and whether the process truly prevents someone from being re-identified are all more complicated than they first appear.
How HIPAA Defines De-Identified Data
In the United States, the HIPAA Privacy Rule sets the legal standard. Under HIPAA, health information is considered de-identified when there is no reasonable basis to believe it could be used, alone or combined with other available information, to identify the person it describes. The rule gives organizations two approved paths to reach that standard.
The first is called the Safe Harbor method. It works like a checklist: remove 18 specific types of identifiers, and confirm that the organization has no actual knowledge the remaining information could identify someone. Those 18 identifiers include the obvious ones (names, phone numbers, email addresses, Social Security numbers) along with less obvious categories:
- Geographic details smaller than a state, including street address, city, county, and most ZIP code digits. The first three digits of a ZIP code can stay only if that three-digit zone covers more than 20,000 people.
- Dates directly tied to an individual, such as birth date, admission date, or discharge date. The year can remain, but the month and day cannot. Ages over 89 must be grouped into a single “90 or older” category.
- Account and device numbers, including medical record numbers, health plan beneficiary numbers, vehicle identifiers, license plates, and device serial numbers.
- Biometric identifiers like fingerprints and voiceprints, full-face photographs, and any other unique identifying number or code.
The second path is the Expert Determination method. Instead of following the checklist, an organization hires a qualified statistical expert who analyzes the data and formally certifies that the risk of re-identification is “very small.” The expert must document their methods and conclusions. This approach is more flexible, because it allows certain identifiers to remain if the overall dataset is structured in a way that makes identification unlikely. It is also more expensive and time-consuming, which is why most organizations default to Safe Harbor.
De-Identified vs. Anonymized vs. Pseudonymized
These three terms often get used interchangeably, but they carry different legal weight depending on where you are.
Under the EU’s General Data Protection Regulation (GDPR), the distinction matters enormously. Anonymized data has been altered so thoroughly that no one, using any means reasonably likely to be available, can link it back to a person. Truly anonymized data is no longer considered personal data at all, and the GDPR stops applying to it entirely.
Pseudonymized data, by contrast, still counts as personal data under the GDPR. Pseudonymization replaces direct identifiers (like a name) with a code or token, but a key exists somewhere that could reverse the process. A hospital might replace patient names with random ID numbers while keeping a separate, secured lookup table. The data is harder to link to individuals, but it is not impossible, so the full weight of privacy law still applies.
De-identification, as HIPAA defines it, sits somewhere between these two. It is more rigorous than pseudonymization but does not necessarily claim that re-identification is impossible, only that the risk is very small. Once data meets HIPAA’s de-identification standard, it is no longer classified as protected health information and can be shared more freely.
Why De-Identified Data Matters
De-identification exists to solve a tension: medical data is enormously valuable for research, public health tracking, and improving care, but sharing it in raw form would expose people’s most private information. Stripping identifiers lets researchers study disease patterns across millions of patients, lets public health agencies monitor outbreaks, and lets technology companies develop diagnostic tools, all without exposing who those patients are.
Insurance companies use de-identified claims data to study cost trends. Pharmaceutical researchers use it to track how medications perform in the real world, outside controlled clinical trials. Public health officials used large de-identified datasets extensively during the COVID-19 pandemic to identify risk factors and evaluate treatments across broad populations. None of that work requires knowing anyone’s name or address.
Re-Identification Risk Is Real
De-identification reduces risk, but it does not eliminate it. As datasets grow larger and more detailed, the possibility that someone could reverse-engineer an identity increases.
A 2024 study published in the Journal of the American Medical Informatics Association demonstrated a straightforward attack on de-identified health datasets. Researchers showed that commonly used encoding strategies kept re-identification rates below 1% for datasets under one million people. But as the dataset scaled up toward the size of national registries (250 million records), re-identification rates climbed to 10% to 20%. The attack worked by combining repeated data patterns with demographic information shared in the clear, like three-digit ZIP codes, to progressively narrow down individual identities.
The core problem is that even after removing the 18 HIPAA identifiers, the remaining data points can form a unique fingerprint. A combination of a rare diagnosis, a specific age, and a general geographic region might describe only one person in the dataset. The more data points that remain, the easier it becomes to single someone out.
Technical Safeguards Beyond Simple Removal
Because stripping identifiers alone has limits, data scientists use additional techniques to make re-identification harder.
One widely used approach is called k-anonymity. The idea is that every record in the dataset should be identical to at least k minus 1 other records on any combination of fields that could serve as indirect identifiers (age, ZIP code, gender, etc.). If k equals 5, then any attempt to look someone up based on those fields will always return at least five matching records, making it impossible to pinpoint which one belongs to the person in question.
K-anonymity has a weakness, though. If all five matching records share the same sensitive value, say the same medical diagnosis, then an attacker learns that information regardless of which record belongs to their target. A refinement called l-diversity addresses this by requiring that each group of matching records contains at least l meaningfully different values for sensitive fields. So instead of five records all showing “diabetes,” the group might include diabetes, asthma, a fracture, depression, and hypertension.
An even stricter standard, t-closeness, requires that the distribution of sensitive values within any group of matching records stays close to the distribution in the overall dataset. This prevents subtler leaks where, for example, a group might technically have diverse diagnoses but be heavily skewed toward one condition compared to the general population.
These techniques involve trade-offs. The more aggressively you generalize or suppress data to protect privacy, the less useful the dataset becomes for research. A dataset where every age is rounded to the nearest decade and every location is reduced to a state loses much of the granularity that makes medical research valuable.
What Stays in De-Identified Data
It is worth understanding what de-identification does not remove. Under the Safe Harbor method, the actual medical content of a record typically remains intact. Diagnoses, lab results, medications prescribed, procedures performed, and treatment outcomes all stay in the dataset. So does the year of any relevant date, the patient’s state of residence, and their age (as long as they are under 90).
This is by design. The goal is to preserve the clinical value of the data while severing its connection to a specific person. A researcher studying heart failure outcomes needs to know what treatments patients received and how they responded. They do not need to know those patients’ names or street addresses.
The practical result is that de-identified health data still contains rich, detailed medical information. It just cannot, in theory, be linked back to you. Whether that theoretical protection holds up as datasets grow larger and analytical tools grow more powerful is an ongoing concern, but for most everyday purposes, de-identification remains the primary mechanism that allows your medical data to contribute to research without your name attached to it.

