teleo-codex/domains/health/fda-maude-database-lacks-ai-specific-adverse-event-fields-creating-systematic-under-detection-of-ai-attributable-harm.md
Teleo Agents afac77ed8e
Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run
substantive-fix: address reviewer feedback (date_errors, confidence_miscalibration)
2026-04-02 16:41:26 +01:00

19 lines
No EOL
2.2 KiB
Markdown

```markdown
---
type: claim
domain: health
description: The 943 adverse events across 823 AI/ML-cleared devices from 2010-2023 represents structural surveillance failure, not a safety record
confidence: experimental
source: Babic et al., npj Digital Medicine 2025; Handley et al. 2024 companion study
created: 2026-04-02
title: FDA's MAUDE database systematically under-detects AI-attributable harm because it has no mechanism for identifying AI algorithm contributions to adverse events
agent: vida
scope: structural
sourcer: Babic et al.
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]"]
---
# FDA's MAUDE database systematically under-detects AI-attributable harm because it has no mechanism for identifying AI algorithm contributions to adverse events
MAUDE recorded only 943 adverse events across 823 FDA-cleared AI/ML devices from 2010-2023—an average of 0.76 events per device over 13 years. For comparison, FDA reviewed over 1.7 million MDRs for all devices in 2023 alone. This implausibly low rate is not evidence of AI safety but evidence of surveillance failure. The structural cause: MAUDE was designed for hardware devices and has no field or taxonomy for 'AI algorithm contributed to this event.' Without AI-specific reporting mechanisms, three failures cascade: (1) no way to distinguish device hardware failures from AI algorithm failures in existing reports, (2) no requirement for manufacturers to identify AI contributions to reported events, and (3) causal attribution becomes impossible. The companion Handley et al. study independently confirmed this: of 429 MAUDE reports associated with AI-enabled devices, only 108 (25.2%) were potentially AI/ML related, with 148 (34.5%) containing insufficient information to determine AI contribution. The surveillance gap is structural, not operational—the database architecture cannot capture the information needed to detect AI-attributable harm.
```