vida: extract claims from 2026-04-21-heudel-ai-deskilling-scoping-review
Some checks are pending
Mirror PR to Forgejo / mirror (pull_request) Waiting to run
Some checks are pending
Mirror PR to Forgejo / mirror (pull_request) Waiting to run
- Source: inbox/queue/2026-04-21-heudel-ai-deskilling-scoping-review.md - Domain: health - Claims: 2, Entities: 0 - Enrichments: 5 - Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5) Pentagon-Agent: Vida <PIPELINE>
This commit is contained in:
parent
02d6c70c96
commit
5a6d1c8821
6 changed files with 73 additions and 38 deletions
|
|
@ -10,21 +10,17 @@ agent: vida
|
|||
scope: causal
|
||||
sourcer: Natali et al.
|
||||
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
|
||||
supports:
|
||||
- "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance'}"
|
||||
- Dopaminergic reinforcement of AI-assisted success creates motivational entrenchment that makes deskilling a behavioral incentive problem, not just a training design problem
|
||||
- "AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance"
|
||||
related:
|
||||
- Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers
|
||||
reweave_edges:
|
||||
- "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-14'}"
|
||||
- Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers|related|2026-04-14
|
||||
- Dopaminergic reinforcement of AI-assisted success creates motivational entrenchment that makes deskilling a behavioral incentive problem, not just a training design problem|supports|2026-04-14
|
||||
- "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-17'}"
|
||||
- "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-18'}"
|
||||
- "AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-19"
|
||||
supports: ["{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance'}", "Dopaminergic reinforcement of AI-assisted success creates motivational entrenchment that makes deskilling a behavioral incentive problem, not just a training design problem", "AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance"]
|
||||
related: ["Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "ai-assistance-produces-neurologically-grounded-irreversible-deskilling-through-prefrontal-disengagement-hippocampal-reduction-and-dopaminergic-reinforcement", "never-skilling-is-structurally-invisible-because-it-lacks-pre-ai-baseline-requiring-prospective-competency-assessment", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling"]
|
||||
reweave_edges: ["{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-14'}", "Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers|related|2026-04-14", "Dopaminergic reinforcement of AI-assisted success creates motivational entrenchment that makes deskilling a behavioral incentive problem, not just a training design problem|supports|2026-04-14", "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-17'}", "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-18'}", "AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-19"]
|
||||
---
|
||||
|
||||
# AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable
|
||||
|
||||
Natali et al.'s systematic review across 10 medical specialties reveals a universal three-phase pattern: (1) AI assistance improves performance metrics while present, (2) extended AI use reduces opportunities for independent skill-building, and (3) performance degrades when AI becomes unavailable, demonstrating dependency rather than augmentation. Quantitative evidence includes: colonoscopy ADR dropping from 28.4% to 22.4% when endoscopists reverted to non-AI procedures after extended AI use (RCT); 30%+ of pathologists reversing correct initial diagnoses when exposed to incorrect AI suggestions under time pressure; 45.5% of ACL diagnosis errors resulting directly from following incorrect AI recommendations across all experience levels. The pattern's consistency across specialties as diverse as neurosurgery, anesthesiology, and geriatrics—not just image-reading specialties—suggests this is a fundamental property of how human cognitive architecture responds to reliable performance assistance, not a specialty-specific implementation problem. The proposed mechanism: AI assistance creates cognitive offloading where clinicians stop engaging prefrontal cortex analytical processes, hippocampal memory formation decreases over repeated exposure, and dopaminergic reinforcement of AI-reliance strengthens, producing skill degradation that becomes visible when AI is removed.
|
||||
|
||||
## Supporting Evidence
|
||||
|
||||
**Source:** Heudel PE et al. 2026, ESMO scoping review
|
||||
|
||||
First comprehensive scoping review (literature through August 2025) confirms consistent deskilling pattern across colonoscopy (6.0pp ADR decline), radiology (12% false-positive increase), pathology (30%+ diagnosis reversals), and cytology (80-85% training volume reduction). Zero studies showed durable skill improvement, making the evidence base one-sided.
|
||||
|
|
|
|||
|
|
@ -10,8 +10,16 @@ agent: vida
|
|||
scope: causal
|
||||
sourcer: Natali et al.
|
||||
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
|
||||
related: ["automation-bias-in-medicine-increases-false-positives-through-anchoring-on-ai-output"]
|
||||
---
|
||||
|
||||
# Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers
|
||||
|
||||
A controlled study of 27 radiologists performing mammography reads found that erroneous AI prompts increased false-positive recalls by up to 12 percentage points, with the effect persisting across experience levels. The mechanism is automation bias: radiologists anchor on AI output rather than conducting fully independent reads, even when they possess the expertise to identify the error. This differs from simple deskilling—it's real-time mis-skilling where the AI's presence actively degrades decision quality below what the clinician would achieve independently. The finding is particularly significant because it occurs in experienced readers, suggesting automation bias is not a training problem but a fundamental feature of human-AI interaction in high-stakes decision contexts. Similar patterns appeared in computational pathology (30%+ diagnosis reversals under time pressure) and ACL diagnosis (45.5% of errors from following incorrect AI recommendations), indicating the mechanism generalizes across imaging modalities and clinical contexts.
|
||||
|
||||
|
||||
## Supporting Evidence
|
||||
|
||||
**Source:** Heudel PE et al. 2026
|
||||
|
||||
Radiology evidence from Heudel review: erroneous AI prompts increased false-positive recalls by up to 12% even among experienced radiologists, demonstrating automation bias operates in expert practitioners, not just novices. This confirms the anchoring mechanism operates across experience levels.
|
||||
|
|
|
|||
|
|
@ -10,26 +10,17 @@ agent: vida
|
|||
scope: causal
|
||||
sourcer: Artificial Intelligence Review (Springer Nature)
|
||||
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
|
||||
supports:
|
||||
- Never-skilling in clinical AI is structurally invisible because it lacks a pre-AI baseline for comparison, requiring prospective competency assessment before AI exposure to detect
|
||||
- "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance'}"
|
||||
- AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable
|
||||
- Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers
|
||||
- Never-skilling — the failure to acquire foundational clinical competencies because AI was present during training — poses a detection-resistant, potentially unrecoverable threat to medical education that is structurally worse than deskilling
|
||||
reweave_edges:
|
||||
- Never-skilling in clinical AI is structurally invisible because it lacks a pre-AI baseline for comparison, requiring prospective competency assessment before AI exposure to detect|supports|2026-04-12
|
||||
- "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-14'}"
|
||||
- AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable|supports|2026-04-14
|
||||
- Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers|supports|2026-04-14
|
||||
- Never-skilling — the failure to acquire foundational clinical competencies because AI was present during training — poses a detection-resistant, potentially unrecoverable threat to medical education that is structurally worse than deskilling|supports|2026-04-14
|
||||
- "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|related|2026-04-17'}"
|
||||
- "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-18'}"
|
||||
- "AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|related|2026-04-19"
|
||||
related:
|
||||
- "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance'}"
|
||||
- "AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance"
|
||||
supports: ["Never-skilling in clinical AI is structurally invisible because it lacks a pre-AI baseline for comparison, requiring prospective competency assessment before AI exposure to detect", "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance'}", "AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable", "Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers", "Never-skilling \u2014 the failure to acquire foundational clinical competencies because AI was present during training \u2014 poses a detection-resistant, potentially unrecoverable threat to medical education that is structurally worse than deskilling"]
|
||||
reweave_edges: ["Never-skilling in clinical AI is structurally invisible because it lacks a pre-AI baseline for comparison, requiring prospective competency assessment before AI exposure to detect|supports|2026-04-12", "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-14'}", "AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable|supports|2026-04-14", "Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers|supports|2026-04-14", "Never-skilling \u2014 the failure to acquire foundational clinical competencies because AI was present during training \u2014 poses a detection-resistant, potentially unrecoverable threat to medical education that is structurally worse than deskilling|supports|2026-04-14", "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|related|2026-04-17'}", "{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|supports|2026-04-18'}", "AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance|related|2026-04-19"]
|
||||
related: ["{'AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms': 'prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance'}", "AI assistance may produce neurologically-grounded, partially irreversible skill degradation through three concurrent mechanisms: prefrontal disengagement, hippocampal memory formation reduction, and dopaminergic reinforcement of AI reliance", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine", "never-skilling-is-structurally-invisible-because-it-lacks-pre-ai-baseline-requiring-prospective-competency-assessment", "ai-assistance-produces-neurologically-grounded-irreversible-deskilling-through-prefrontal-disengagement-hippocampal-reduction-and-dopaminergic-reinforcement"]
|
||||
---
|
||||
|
||||
# Clinical AI introduces three distinct skill failure modes — deskilling (existing expertise lost through disuse), mis-skilling (AI errors adopted as correct), and never-skilling (foundational competence never acquired) — requiring distinct mitigation strategies for each
|
||||
|
||||
This systematic review identifies three mechanistically distinct pathways through which clinical AI degrades physician competence. **Deskilling** occurs when existing expertise atrophies through disuse: colonoscopy polyp detection dropped from 28.4% to 22.4% after 3 months of AI use, and experienced radiologists showed 12% increased false-positive recalls after exposure to erroneous AI prompts. **Mis-skilling** occurs when clinicians actively learn incorrect patterns from systematically biased AI outputs: in computational pathology studies, 30%+ of participants reversed correct initial diagnoses after exposure to incorrect AI suggestions under time constraints. **Never-skilling** is categorically different: trainees who begin clinical education with AI assistance may never develop foundational competencies. Junior radiologists are far less likely than senior colleagues to detect AI errors — not because they've lost skills, but because they never acquired them. This is structurally invisible because there's no pre-AI baseline to compare against. The review documents mitigation strategies including AI-off drills, structured assessment pre-AI review, and curriculum redesign with explicit competency development before AI exposure. The key insight is that these three failure modes require fundamentally different interventions: deskilling requires practice maintenance, mis-skilling requires error detection training, and never-skilling requires prospective competency assessment before AI exposure.
|
||||
|
||||
## Extending Evidence
|
||||
|
||||
**Source:** Heudel PE et al. 2026, UK cervical screening consolidation
|
||||
|
||||
UK cytology lab consolidation provides first structural never-skilling mechanism: 80-85% training volume reduction through consolidation from 45 to 8 labs. This extends the never-skilling concept from individual cognitive failure to institutional infrastructure destruction. The mechanism is not 'physicians never learn because AI does it for them' but 'training infrastructure is dismantled so learning becomes impossible.'
|
||||
|
|
|
|||
|
|
@ -0,0 +1,18 @@
|
|||
---
|
||||
type: claim
|
||||
domain: health
|
||||
description: UK cervical screening AI deployment consolidated labs from 45 to 8 centers, reducing training case volumes by 80-85 percent and structurally eliminating the apprenticeship infrastructure needed to acquire diagnostic skills
|
||||
confidence: experimental
|
||||
source: "Heudel et al. 2026, ESMO Real World Data & Digital Oncology scoping review"
|
||||
created: 2026-04-21
|
||||
title: Cytology lab consolidation creates never-skilling pathway through 80 percent training volume destruction
|
||||
agent: vida
|
||||
scope: structural
|
||||
sourcer: Heudel PE, Crochet H, Filori Q, Bachelot T, Blay JY
|
||||
supports: ["clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling"]
|
||||
related: ["human-in-the-loop-clinical-ai-degrades-to-worse-than-ai-alone-because-physicians-both-de-skill-from-reliance-and-introduce-errors-when-overriding-correct-outputs", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling", "never-skilling-is-structurally-invisible-because-it-lacks-pre-ai-baseline-requiring-prospective-competency-assessment"]
|
||||
---
|
||||
|
||||
# Cytology lab consolidation creates never-skilling pathway through 80 percent training volume destruction
|
||||
|
||||
Following UK cervical screening consolidation with AI-assisted reading, case volumes reduced 80-85% while labs consolidated from 45 to 8 centers. The authors identify this as having 'major implications for training capacity.' This represents a distinct mechanism from individual cognitive deskilling: the training system itself is structurally dismantled. When training volume is eliminated at this scale, clinicians never acquire the skill in the first place — the never-skilling pathway. This is worse than deskilling because it's irreversible without rebuilding training infrastructure. The mechanism is structural volume destruction, not individual cognitive dependency. Unlike deskilling (where physicians forget skills they once had) or misskilling (where AI prompts cause real-time errors), never-skilling operates at the institutional level by destroying the apprenticeship pipeline. This finding extends the existing KB's three-failure-mode framework (deskilling, misskilling, never-skilling) with the first documented case of structural never-skilling through lab consolidation.
|
||||
|
|
@ -10,12 +10,16 @@ agent: vida
|
|||
scope: causal
|
||||
sourcer: Journal of Experimental Orthopaedics / Wiley
|
||||
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
|
||||
related:
|
||||
- AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable
|
||||
reweave_edges:
|
||||
- AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable|related|2026-04-14
|
||||
related: ["AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling", "never-skilling-is-structurally-invisible-because-it-lacks-pre-ai-baseline-requiring-prospective-competency-assessment", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "ai-assistance-produces-neurologically-grounded-irreversible-deskilling-through-prefrontal-disengagement-hippocampal-reduction-and-dopaminergic-reinforcement"]
|
||||
reweave_edges: ["AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable|related|2026-04-14"]
|
||||
---
|
||||
|
||||
# Never-skilling — the failure to acquire foundational clinical competencies because AI was present during training — poses a detection-resistant, potentially unrecoverable threat to medical education that is structurally worse than deskilling
|
||||
|
||||
Never-skilling is formally defined in peer-reviewed literature as distinct from and more dangerous than deskilling for three structural reasons. First, it is unrecoverable: deskilling allows clinicians to re-engage practice and rebuild atrophied skills, but never-skilling means foundational representations were never formed — there is nothing to rebuild from. Second, it is detection-resistant: clinicians who never developed skills don't know what they're missing, and supervisors reviewing AI-assisted work cannot distinguish never-skilled from skilled performance. Third, it is prospectively invisible: the harm manifests 5-10 years after training when current trainees become independent practitioners, creating a delayed-onset safety crisis. The JEO review explicitly states 'never-skilling poses a greater long-term threat to medical education than deskilling' because early reliance on automation prevents acquisition of foundational clinical reasoning and procedural competencies. Supporting evidence includes findings that more than one-third of advanced medical students failed to identify erroneous LLM answers to clinical scenarios, and significant negative correlation between frequent AI tool use and critical thinking abilities. The concept has graduated from informal commentary to formal peer-reviewed definition across NEJM, JEO, and Lancet Digital Health, though no prospective RCT yet exists comparing AI-naive versus AI-exposed-from-training cohorts on downstream clinical performance.
|
||||
|
||||
## Supporting Evidence
|
||||
|
||||
**Source:** Heudel PE et al. 2026
|
||||
|
||||
Cytology lab consolidation demonstrates unrecoverability: 37 labs closed (45 to 8), 80-85% training volume eliminated. Reversing this requires rebuilding physical infrastructure, not just retraining individuals. This confirms never-skilling is structurally worse than deskilling because the recovery path requires institutional reconstruction.
|
||||
|
|
|
|||
|
|
@ -0,0 +1,18 @@
|
|||
---
|
||||
type: claim
|
||||
domain: health
|
||||
description: Comprehensive scoping review through August 2025 found consistent evidence of AI-induced deskilling across four specialties but zero studies demonstrating lasting skill improvement after AI exposure
|
||||
confidence: likely
|
||||
source: Heudel et al. 2026 scoping review, literature through August 2025
|
||||
created: 2026-04-21
|
||||
title: No peer-reviewed evidence of durable physician upskilling from AI exposure as of mid-2026
|
||||
agent: vida
|
||||
scope: correlational
|
||||
sourcer: Heudel PE, Crochet H, Filori Q, Bachelot T, Blay JY
|
||||
supports: ["human-in-the-loop-clinical-ai-degrades-to-worse-than-ai-alone-because-physicians-both-de-skill-from-reliance-and-introduce-errors-when-overriding-correct-outputs", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine"]
|
||||
related: ["human-in-the-loop-clinical-ai-degrades-to-worse-than-ai-alone-because-physicians-both-de-skill-from-reliance-and-introduce-errors-when-overriding-correct-outputs", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine", "automation-bias-in-medicine-increases-false-positives-through-anchoring-on-ai-output", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling", "never-skilling-is-structurally-invisible-because-it-lacks-pre-ai-baseline-requiring-prospective-competency-assessment", "human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs"]
|
||||
---
|
||||
|
||||
# No peer-reviewed evidence of durable physician upskilling from AI exposure as of mid-2026
|
||||
|
||||
The Heudel et al. scoping review examined literature through August 2025 across colonoscopy, radiology, pathology, and cytology. Authors conclude: 'empirical studies consistently demonstrate that AI can inadvertently impair physicians' performance.' The review found NO opposing evidence — no studies showed durable improvement in physician skills after AI exposure. This null result is itself significant: after 5+ years of clinical AI deployment, there is no peer-reviewed evidence of durable skill improvement. The authors searched for counter-evidence and found none. This creates a lopsided evidence base: strong consistent evidence of deskilling (colonoscopy ADR dropped 6.0 percentage points when AI removed; radiology false-positive recalls increased 12% from erroneous AI prompts; pathology showed 30%+ diagnosis reversals from incorrect AI suggestions) versus zero evidence of lasting upskilling. The absence of upskilling evidence is notable because it contradicts the common assumption that AI 'calibrates' or 'teaches' clinicians. If such effects existed and were durable, they should be detectable in the literature by now.
|
||||
Loading…
Reference in a new issue