| claim |
health |
FDA expanded CDS enforcement discretion on January 6 2026 in the same month ECRI published AI chatbots as the number one health technology hazard revealing temporal contradiction between regulatory rollback and patient safety alarm |
experimental |
FDA CDS Guidance January 2026, ECRI 2026 Health Technology Hazards Report |
2026-04-02 |
Clinical AI deregulation is occurring during active harm accumulation not after evidence of safety as demonstrated by simultaneous FDA enforcement discretion expansion and ECRI top hazard designation in January 2026 |
vida |
structural |
ECRI |
|
| Clinical AI chatbot misuse is a documented ongoing harm source not a theoretical risk as evidenced by ECRI ranking it the number one health technology hazard for two consecutive years |
| FDA's 2026 CDS guidance expands enforcement discretion to cover AI tools providing single clinically appropriate recommendations while leaving clinical appropriateness undefined and requiring no bias evaluation or post-market surveillance |
| The clinical AI safety gap is doubly structural |
| FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm |
|
| Regulatory rollback of clinical AI oversight in EU and US during 2025-2026 represents coordinated or parallel regulatory capture occurring simultaneously with accumulating research evidence of failure modes |
| Regulatory vacuum emerges when deregulation outpaces safety evidence accumulation creating institutional epistemic divergence between regulators and health authorities |
|
| Clinical AI chatbot misuse is a documented ongoing harm source not a theoretical risk as evidenced by ECRI ranking it the number one health technology hazard for two consecutive years|supports|2026-04-03 |
| FDA's 2026 CDS guidance expands enforcement discretion to cover AI tools providing single clinically appropriate recommendations while leaving clinical appropriateness undefined and requiring no bias evaluation or post-market surveillance|supports|2026-04-03 |
| The clinical AI safety gap is doubly structural |
| FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm|supports|2026-04-07 |
|
| Regulatory rollback of clinical AI oversight in EU and US during 2025-2026 represents coordinated or parallel regulatory capture occurring simultaneously with accumulating research evidence of failure modes|supports|2026-04-07 |
| Regulatory vacuum emerges when deregulation outpaces safety evidence accumulation creating institutional epistemic divergence between regulators and health authorities|supports|2026-04-07 |
| All three major clinical AI regulatory tracks converged on adoption acceleration rather than safety evaluation in Q1 2026|related|2026-04-07 |
|
| All three major clinical AI regulatory tracks converged on adoption acceleration rather than safety evaluation in Q1 2026 |
|