MedTech Compliance Architecture: Why Compliant Devices Still Fail in the Field
Key takeaways
- Class I recalls hit 15-year high in 2024 with impacted units up 55% YoY; device failure overtook manufacturing defects as leading cause for first time
- Software design errors drive ~50% of AI/ML device recalls (Chen et al., 27 years FDA data); most failures emerge when architectures can’t constrain behavior under change
- Regulated boundary now includes cloud services, update servers, data pipelines, ML lifecycle—but compliance responsibility still localized to embedded teams creates structural gap
- Legacy devices lack SBOMs and secure updates; October 2025 enforcement may refuse submissions with missing SBOM data, forcing costly architectural retrofits
Compliance Was Always Required. Reality Changed Anyway
Engineers working under IEC 62304 and ISO 13485 already know that compliance must be addressed at the architecture stage. That principle has been true for years. This article is not a reminder.
What changed is that the architecture most teams review no longer matches the system regulators evaluate once the device is in the field.
That gap explains why recall patterns continue to worsen even as organizations invest more heavily in process maturity, tooling, and documentation.
In 2024, IQVIA recorded more than 16,000 medical device recall and corrective action events worldwide, affecting over 70 million devices. Software-driven and connected products made up a growing share. At the same time, Class I recalls reached a fifteen-year high, with impacted units increasing more than fifty percent year over year. Failures are not only more frequent. They are more severe.
A 2025 longitudinal analysis by Chen and colleagues, covering 27 years of FDA recall data for AI- and ML-enabled medical devices, found that design and development causes account for roughly half of recalls in this segment. Software design errors were the most frequent root cause.
This gap shows up for CTOs, system architects, and engineering leaders responsible for medical device software that spans device firmware, cloud platforms, update infrastructure, and data pipelines.
Systems changed faster than the architectures and organizations built to control them, and recalls followed.
Compliance by architecture is no longer about proving that design controls existed at project start. It is about whether the system can absorb updates, cybersecurity events, data drift, integration pressure, and post-market change without breaking its own safety case.
What Actually Changed Since ‘We Already Knew This’
The standards did not move. The system did.
In modern medical device software, compliance failures increasingly originate in system architecture rather than implementation defects.
Regulators now routinely treat cloud services, update servers, telemetry pipelines, data handling, and ML lifecycle elements as part of the regulated system when they influence device behavior. This is no longer edge guidance. It is how inspections, recalls, and remediation are handled in practice.
At the same time, software stopped being static. Devices are expected to receive updates. Cybersecurity vulnerabilities force change. ML models evolve. Integration environments shift. Each change stresses assumptions that were safe when releases were infrequent and tightly bound.
Cybersecurity guidance made this unavoidable. Since late 2023, connected-device submissions that lack complete, machine-readable SBOMs routinely trigger additional information requests that stretch clearance timelines. Starting in 2025, missing SBOM data may result in a refusal to accept submissions altogether. A system that cannot account for its software composition cannot be safely maintained, regardless of how well it passed initial verification.
Nothing here contradicts IEC 62304. What it breaks is the assumption that architecture can be reviewed once and then left alone while change is handled procedurally.
Most Software-Related Recalls Are Architecture Failures
Many recalls categorized as software-related are not caused by isolated defects. They emerge when architectures fail to constrain system behavior under change.
The pattern is familiar. A change is introduced to improve performance, security, usability, or data handling. The change itself is small. The system response is not. There is no safe rollback across device, application, and cloud. Configuration changes bypass formal change control because they live outside firmware repositories. UI workflows that passed summative testing behave differently under real clinical pressure. Dependency updates invalidate prior verification assumptions.
This shift is now visible in recall categorization itself. In 2024, device failure overtook manufacturing defects as the leading cause of medical device recalls, reflecting a move away from execution errors toward system-level design and behavior failures.
Chen’s analysis shows that in AI- and ML-enabled devices, software design and change handling dominate recall causes. Independent reviews of openFDA data reach the same conclusion once recall narratives are examined closely. UI-driven use errors, often misclassified as generic software issues, account for a large share of serious events.
These failures are rarely caused by a lack of effort or awareness. They result from architectures that allow unsafe states to exist and unsafe changes to propagate.
Why Regulated Behavior Is No Longer Owned by Embedded Teams
Most organizations still localize compliance responsibility to embedded engineering and quality teams. That model reflected how products were built a decade ago.
It no longer reflects how regulated behavior is created.
Today, device behavior is shaped by cloud services that control logic and updates, by DevOps pipelines that decide what runs in the field, by data pipelines that influence inference and thresholds, and by security teams responding to vulnerabilities under time pressure. These teams often operate under different governance models, incentives, and review practices, even though their decisions directly affect safety and compliance.
The result is a structural gap. Safety-critical design decisions are made outside the parts of the organization trained to recognize them as such.
A simple question exposes the issue. Who owns change control for updating server configuration? Not the firmware. The update server. If ownership is unclear, the system carries latent compliance risk even if audits pass today.
Where Compliance Architecture Breaks First
Teams adding ML and data-driven features are exposed because the risk does not live in the model itself. It lives in the architecture around it. Training data, thresholds, deployment paths, monitoring, and planned changes now form part of the safety argument. Treating these elements as data science concerns rather than safety-critical system components creates blind spots that traditional test plans never exercise.
Organizations accelerating development through AI-assisted or low-code tools face a related issue. The tools increase throughput, but they also compress decision cycles. Regulators do not distinguish between human-written and AI-generated code. They expect traceable requirements, controlled design decisions, verified behavior, and documented releases. Without an architecture that enforces lifecycle discipline through tooling and pipelines, velocity outpaces understanding.
Legacy programs are the most exposed and the most common. Devices designed three to five years ago were rarely built with continuous updates, observability, and dependency control in mind. Many lack SBOMs tied to releases or secure update paths. Retrofitting these capabilities often requires architectural rework across devices, gateways, and integrations. In many cases, the cost exceeds building a compliance-ready platform from scratch.
How to Tell If Your Architecture Will Survive Change
Most teams can show that the architecture was reviewed. That is not where programs fail.
They fail when the system is forced to change.
A useful compliance architecture check does not ask whether standards were considered. It asks whether the architecture still holds under pressure.
Can you reproduce what is in the field, including dependencies, cloud configuration, update logic, and model versions? Can you understand the impact of a change before it ships, across hazards, requirements, tests, and deployed components? Can you roll back safely across firmware, applications, cloud services, configuration, and models? Is compliance enforced by tooling or by people remembering steps? Is ownership aligned with system boundaries rather than organizational convenience?
When teams struggle to answer these questions clearly, recalls usually follow within one or two release cycles.
The Cost of Treating Architecture as ‘Later’
Non-routine quality events cost the MedTech industry billions annually, but the more damaging cost is structural drag.
Weak architecture turns every change into a negotiation with the past. Revalidation cycles slow releases. Updates are delayed or avoided. Features are removed to reduce risk. After major recalls, organizations often freeze innovation while competitors move.
Regulatory posture reflects this shift. In 2025, the FDA expanded its Communications Pilot to Enhance the Medical Device Recall Program to cover all medical device categories, signalling that recall risk and communication are now treated as systemic concerns across the industry, not edge cases.
The cost is not limited to remediation. It is a loss of momentum and strategic optionality.
Architecture Is Now a Compliance Decision
This is not a reminder that architecture matters. Engineers already know that.
It is an acknowledgement that modern MedTech systems have outgrown the compliance mental model that many organizations still apply. Cloud services, frequent updates, data pipelines, and cybersecurity turned architecture into a continuous compliance surface.
You will pay for that reality either through deliberate system design or through unplanned remediation under regulatory pressure.
The difference is not whether the cost appears.
It is whether you control when and how it does.

