vida: extract claims from 2026-04-22-pmc11780016-radiology-ai-upskilling-study-2025 #3771

Closed
vida wants to merge 1 commit from extract/2026-04-22-pmc11780016-radiology-ai-upskilling-study-2025-e6a1 into main
2 changed files with 29 additions and 40 deletions

View file

@ -1,44 +1,19 @@
--- ---
agent: vida
confidence: likely
created: 2026-04-13
description: Systematic review across 10 medical specialties (radiology, neurosurgery, anesthesiology, oncology, cardiology, pathology, fertility medicine, geriatrics, psychiatry, ophthalmology) finds universal
pattern of skill degradation following AI removal
domain: health
related:
- Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers
- ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine
- clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling
- ai-assistance-produces-neurologically-grounded-irreversible-deskilling-through-prefrontal-disengagement-hippocampal-reduction-and-dopaminergic-reinforcement
- never-skilling-is-structurally-invisible-because-it-lacks-pre-ai-baseline-requiring-prospective-competency-assessment
- never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling
- economic forces push humans out of every cognitive loop where output quality is independently verifiable because human-in-the-loop is a cost that competitive markets eliminate
related_claims:
- '[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]'
reweave_edges:
- '{''AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms'': ''prefrontal disengagement, hippocampal memory formation reduction,
and dopaminergic reinforcement of AI reliance|supports|2026-04-14''}'
- Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers|related|2026-04-14
- Dopaminergic reinforcement of AI-assisted success creates motivational entrenchment that makes deskilling a behavioral incentive problem, not just a training design problem|supports|2026-04-14
- '{''AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms'': ''prefrontal disengagement, hippocampal memory formation reduction,
and dopaminergic reinforcement of AI reliance|supports|2026-04-17''}'
- '{''AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms'': ''prefrontal disengagement, hippocampal memory formation reduction,
and dopaminergic reinforcement of AI reliance|supports|2026-04-18''}'
- 'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and
dopaminergic reinforcement of AI reliance|supports|2026-04-19'
scope: causal
source: Natali et al., Artificial Intelligence Review 2025, mixed-method systematic review
sourced_from:
- inbox/archive/health/2026-04-13-natali-2025-ai-deskilling-comprehensive-review.md
sourcer: Natali et al.
supports:
- '{''AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms'': ''prefrontal disengagement, hippocampal memory formation reduction,
and dopaminergic reinforcement of AI reliance''}'
- Dopaminergic reinforcement of AI-assisted success creates motivational entrenchment that makes deskilling a behavioral incentive problem, not just a training design problem
- 'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and
dopaminergic reinforcement of AI reliance'
title: AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable
type: claim type: claim
domain: health
description: Systematic review across 10 medical specialties (radiology, neurosurgery, anesthesiology, oncology, cardiology, pathology, fertility medicine, geriatrics, psychiatry, ophthalmology) finds universal pattern of skill degradation following AI removal
confidence: likely
source: Natali et al., Artificial Intelligence Review 2025, mixed-method systematic review
created: 2026-04-13
agent: vida
related: ["Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "ai-assistance-produces-neurologically-grounded-irreversible-deskilling-through-prefrontal-disengagement-hippocampal-reduction-and-dopaminergic-reinforcement", "never-skilling-is-structurally-invisible-because-it-lacks-pre-ai-baseline-requiring-prospective-competency-assessment", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling", "economic forces push humans out of every cognitive loop where output quality is independently verifiable because human-in-the-loop is a cost that competitive markets eliminate", "no-peer-reviewed-evidence-of-durable-physician-upskilling-from-ai-exposure-as-of-mid-2026"]
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
reweave_edges: ["{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-14'}", "Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers|related|2026-04-14", "Dopaminergic reinforcement of AI-assisted success creates motivational entrenchment that makes deskilling a behavioral incentive problem, not just a training design problem|supports|2026-04-14", "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-17'}", "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-18'}", "AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-19"]
scope: causal
sourced_from: ["inbox/archive/health/2026-04-13-natali-2025-ai-deskilling-comprehensive-review.md"]
sourcer: Natali et al.
supports: ["{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance'}", "Dopaminergic reinforcement of AI-assisted success creates motivational entrenchment that makes deskilling a behavioral incentive problem, not just a training design problem", "AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance"]
title: AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable
--- ---
# AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable # AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable
@ -50,3 +25,10 @@ Natali et al.'s systematic review across 10 medical specialties reveals a univer
**Source:** Heudel PE et al. 2026, ESMO scoping review **Source:** Heudel PE et al. 2026, ESMO scoping review
First comprehensive scoping review (literature through August 2025) confirms consistent deskilling pattern across colonoscopy (6.0pp ADR decline), radiology (12% false-positive increase), pathology (30%+ diagnosis reversals), and cytology (80-85% training volume reduction). Zero studies showed durable skill improvement, making the evidence base one-sided. First comprehensive scoping review (literature through August 2025) confirms consistent deskilling pattern across colonoscopy (6.0pp ADR decline), radiology (12% false-positive increase), pathology (30%+ diagnosis reversals), and cytology (80-85% training volume reduction). Zero studies showed durable skill improvement, making the evidence base one-sided.
## Challenging Evidence
**Source:** Heudel et al., Insights into Imaging, Jan 2025
Heudel et al. radiology study shows residents can detect large AI errors (>3 points) while maintaining lower error rates (2.75-2.88) than the AI's major mistakes, suggesting some preservation of critical judgment even during AI-assisted work. This 'resilience to AI errors above acceptability threshold' indicates the deskilling pattern may not be uniform - physicians may retain ability to catch egregious errors while still experiencing skill degradation in more subtle diagnostic tasks. However, n=8 sample size and lack of longitudinal follow-up limit generalizability.

View file

@ -8,7 +8,7 @@ secondary_domains: ["ai-alignment", "collective-intelligence"]
title: Does human oversight improve or degrade AI clinical decision-making? title: Does human oversight improve or degrade AI clinical decision-making?
claims: ["human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs.md", "AI middleware bridges consumer wearable data to clinical utility because continuous data is too voluminous for direct clinician review.md"] claims: ["human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs.md", "AI middleware bridges consumer wearable data to clinical utility because continuous data is too voluminous for direct clinician review.md"]
surfaced_by: leo surfaced_by: leo
related: ["divergence-human-ai-clinical-collaboration-enhance-or-degrade", "the physician role shifts from information processor to relationship manager as AI automates documentation triage and evidence synthesis", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine", "medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling"] related: ["divergence-human-ai-clinical-collaboration-enhance-or-degrade", "the physician role shifts from information processor to relationship manager as AI automates documentation triage and evidence synthesis", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine", "medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "no-peer-reviewed-evidence-of-durable-physician-upskilling-from-ai-exposure-as-of-mid-2026"]
--- ---
# Does human oversight improve or degrade AI clinical decision-making? # Does human oversight improve or degrade AI clinical decision-making?
@ -62,3 +62,10 @@ Topics:
**Source:** Oettl et al. 2026, Journal of Experimental Orthopaedics PMC12955832 **Source:** Oettl et al. 2026, Journal of Experimental Orthopaedics PMC12955832
Oettl et al. 2026 provides the strongest articulation of the upskilling thesis, arguing that AI creates 'micro-learning at point of care' through review-confirm-override loops. However, the paper's own evidence base consists entirely of 'performance with AI present' studies (Heudel et al. showing 22% higher inter-rater agreement, COVID-19 detection achieving near-perfect accuracy with AI). No cited studies measure durable skill retention after AI training in a no-AI follow-up arm. The paper explicitly acknowledges: 'deskilling threat is real if trainees never develop foundational competencies' and 'further studies needed on surgical AI's long-term patient outcomes.' This represents the upskilling hypothesis at its strongest—and reveals that even its strongest proponents lack prospective longitudinal evidence. Oettl et al. 2026 provides the strongest articulation of the upskilling thesis, arguing that AI creates 'micro-learning at point of care' through review-confirm-override loops. However, the paper's own evidence base consists entirely of 'performance with AI present' studies (Heudel et al. showing 22% higher inter-rater agreement, COVID-19 detection achieving near-perfect accuracy with AI). No cited studies measure durable skill retention after AI training in a no-AI follow-up arm. The paper explicitly acknowledges: 'deskilling threat is real if trainees never develop foundational competencies' and 'further studies needed on surgical AI's long-term patient outcomes.' This represents the upskilling hypothesis at its strongest—and reveals that even its strongest proponents lack prospective longitudinal evidence.
## Extending Evidence
**Source:** Heudel et al., Insights into Imaging, Jan 2025 (PMC11780016)
Heudel et al. (2025) radiology study (n=8 residents, 150 chest X-rays) shows 22% improvement in inter-rater agreement (ICC-1: 0.665→0.813) and significant error reduction (p<0.001) WITH AI present. However, study design lacks post-training no-AI assessment, so it documents performance improvement during AI use, not durable skill retention. Residents showed 'resilience to AI errors above acceptability threshold' - when AI made major errors (>3 points), residents maintained average errors around 2.75-2.88, suggesting ability to detect large AI mistakes. Critical limitation: this is the primary study cited as evidence for AI 'upskilling' of physicians, but it only measures performance with AI present, not whether skills persist after AI removal. Contrast with colonoscopy RCT showing ADR decline from 28.4%→22.4% when AI removed, which tests the durable skill question this study leaves unanswered.