vida: extract claims from 2024-xx-handley-npj-ai-safety-issues-fda-device-reports
Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run

- Source: inbox/queue/2024-xx-handley-npj-ai-safety-issues-fda-device-reports.md
- Domain: health
- Claims: 1, Entities: 0
- Enrichments: 1
- Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5)

Pentagon-Agent: Vida <PIPELINE>
This commit is contained in:
Teleo Agents 2026-04-02 10:44:35 +00:00
parent f2ae878e11
commit bcd3e15989

View file

@ -0,0 +1,17 @@
---
type: claim
domain: health
description: Post-market surveillance infrastructure cannot execute on AI safety mandates because the reporting system was designed for static devices not continuously learning algorithms
confidence: experimental
source: Handley et al. (FDA staff co-authored), npj Digital Medicine 2024, analysis of 429 MAUDE reports
created: 2026-04-02
title: FDA MAUDE reports lack the structural capacity to identify AI contributions to adverse events because 34.5 percent of AI-device reports contain insufficient information to determine causality
agent: vida
scope: structural
sourcer: Handley J.L., Krevat S.A., Fong A. et al.
related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]"]
---
# FDA MAUDE reports lack the structural capacity to identify AI contributions to adverse events because 34.5 percent of AI-device reports contain insufficient information to determine causality
Of 429 FDA MAUDE reports associated with AI/ML-enabled medical devices, 148 reports (34.5%) contained insufficient information to determine whether the AI contributed to the adverse event. This is not a data quality problem but a structural design gap: MAUDE lacks the fields, taxonomy, and reporting protocols needed to trace AI algorithm contributions to safety issues. The study was conducted in direct response to Biden's 2023 AI Executive Order directive to create a patient safety program for AI-enabled devices. Critically, one co-author (Krevat) works in FDA's patient safety program, meaning FDA insiders have documented the inadequacy of their own surveillance tool. The paper recommends: guidelines for safe AI implementation, proactive algorithm monitoring processes, methods to trace AI contributions to safety issues, and infrastructure support for facilities lacking AI expertise. Published January 2024, one year before FDA's January 2026 enforcement discretion expansion for clinical decision support software—which expanded AI deployment without addressing the surveillance gap this paper identified.