A DFMEA, or Design Failure Mode and Effects Analysis, is a structured method engineers use to find potential ways a product design could fail before it ever reaches production. The goal is straightforward: identify what could go wrong, assess how serious each risk is, and fix the most dangerous problems while the design is still on paper. It’s one of the most widely used risk assessment tools in automotive, aerospace, medical device, and electronics engineering.
How DFMEA Fits Into Product Development
A DFMEA focuses specifically on the product’s design and functionality. It asks questions like: Could this material crack under stress? Could this circuit overheat in normal use? Could this component fail in a way that injures someone? The analysis happens during the design phase, ideally early enough that changes are still cheap and practical to make.
This is different from a Process FMEA (PFMEA), which looks at what could go wrong during manufacturing or assembly. A DFMEA might flag that a shaft connection can’t transmit enough torque because of a design flaw in the spline specifications. A PFMEA would flag that a machine on the factory floor might install that shaft incorrectly. One targets design integrity, the other targets production workflows. Both feed into the broader quality system, but they analyze completely different types of risk.
The 7 Steps of a DFMEA
The current industry standard follows a seven-step framework developed jointly by the Automotive Industry Action Group (AIAG) and the German Association of the Automotive Industry (VDA). While it originated in automotive, the same structure applies across industries.
- Planning and preparation. Define the scope of the analysis: which system, subsystem, or component you’re evaluating, and what standards or customer requirements apply.
- Structure analysis. Break the design down into a hierarchy. For a hybrid vehicle drivetrain, that might mean going from the full system down to the inverter, then to individual wiring connections and communication buses.
- Function analysis. Identify what each element is supposed to do. A motor connection needs to transmit a specific amount of torque. An inverter needs to send timeout messages when communication drops.
- Failure analysis. For each function, list the ways it could fail (failure modes), what would happen if it did (effects), and what might cause it (root causes). A non-concentric connection might fail to transmit torque because of improper alignment or incorrect spline specifications.
- Risk analysis. Score each failure for severity, likelihood, and detectability (more on this below).
- Optimization. Assign corrective actions to the highest-priority risks. This could mean redesigning a component, adding a tolerance analysis, or introducing new testing protocols.
- Communication. Share results with stakeholders, link findings to control plans, and document everything for audits and future reference.
How Risks Are Scored
Every failure mode gets rated on three scales, each from 1 to 10:
Severity (S) measures how bad the consequences would be if the failure actually happened. A score of 1 means the effect is insignificant. A 10 means it’s catastrophic, typically involving a safety hazard or regulatory violation. Severity is based on the worst-case effect on the end user.
Occurrence (O) estimates how likely the failure is to happen. A 1 means it’s extremely unlikely. A 10 means it’s practically inevitable given the current design. This rating draws on historical data, engineering judgment, and testing results from similar products.
Detection (D) rates how well your current design controls can catch the problem before it reaches the customer. A 1 means detection is virtually certain. A 10 means there’s no control in place and the failure will go unnoticed. This scale runs counterintuitively: lower numbers are better.
RPN vs. Action Priority
The traditional approach multiplies all three scores together to produce a Risk Priority Number (RPN). A failure with severity 8, occurrence 4, and detection 6 gets an RPN of 192. Teams then ranked failures by RPN and worked down the list.
The problem with RPN is that it treats all three factors equally. A failure scoring 9 for severity and 2 for everything else (RPN of 36) would rank lower than one scoring 4 across all three (RPN of 64), even though the first failure is far more dangerous. The newer Action Priority (AP) method, introduced in the AIAG-VDA handbook, addresses this by weighting severity first, then occurrence, then detection. Instead of a single number, it assigns failures to priority levels (high, medium, low) that better reflect actual risk.
What a Real DFMEA Looks Like
In practice, a DFMEA is a large spreadsheet or database where each row captures one failure scenario in detail. To make this concrete, here’s what entries look like from an analysis of a hybrid vehicle drivetrain:
One failure mode identified was a shaft connection that couldn’t transmit the driver’s demanded torque. The root cause was improper alignment combined with incorrect spline specifications for the rated load. No preventive controls existed at the time. The only detection method was visual inspection during vehicle testing, which is late in the process and unreliable for internal alignment issues. This type of entry would score high on severity (loss of drive torque is a safety concern) and high on detection difficulty, making it a top priority for redesign.
Another entry addressed an inverter that failed to send a communication timeout message when it should have. The cause was an overloaded data bus carrying too many signals. The prevention control was straightforward: move the critical torque signal to a separate bus so it wouldn’t get crowded out. Detection relied on monitoring software that could flag when signals were being lost. This is a good example of a failure where a simple architectural change eliminates the root cause entirely.
A third failure had the same symptom (no timeout message) but a different cause: faulty wiring between two controllers. Prevention here relied on a second engineer inspecting the connections, and detection used software verification of the communication link. Same failure mode, different causes, different controls. That’s a key feature of DFMEA: it forces you to trace each failure back to its specific root cause rather than treating symptoms as a lump.
Who Participates in a DFMEA
A DFMEA is never a solo exercise. It requires a cross-functional team because no single engineer has visibility into every aspect of how a design might fail. A typical core team includes a design engineer who owns the DFMEA document, a quality engineer who ensures alignment with industry standards and brings lessons learned from past warranty claims, a system engineer who understands how the component interacts with the larger product, and a safety or compliance expert who flags regulatory concerns.
An FMEA moderator facilitates each session. Their job is to keep discussions structured and focused on risk prevention rather than letting the meeting devolve into form-filling. Depending on the product, the team might also include supplier quality engineers who understand purchased component risks, test engineers who can provide validation data, software engineers for anything involving embedded code, and sometimes a customer representative who can communicate specific performance expectations.
For example, a DFMEA on an electric power steering system might include the design lead defining motor torque functions, a software engineer analyzing signal processing failures, a safety engineer assessing impacts under the automotive functional safety standard ISO 26262, and a supplier quality engineer raising concerns about sensor reliability from a specific vendor.
Industry Standards That Govern DFMEA
The most widely referenced standard is SAE J1739, last updated in 2021. It covers Design FMEA, Process FMEA, and a supplemental type called FMEA-MSR (monitoring and system response). The standard provides required terms, rating charts, and worksheet formats. Compliance is mandatory in many automotive supply chains, and deviations require documented rationale and customer agreement to survive an audit.
In automotive specifically, DFMEA is a core part of the Advanced Product Quality Planning (APQP) process and ties directly into IATF 16949 certification. Medical device manufacturers follow a parallel approach under ISO 14971 for risk management. The underlying logic is the same across industries: systematically identify what could fail, quantify the risk, and prove you’ve addressed it before production begins.

