vida: extract claims from 2026-04-22-oettl-2026-ai-deskilling-to-upskilling-orthopedics
- Source: inbox/queue/2026-04-22-oettl-2026-ai-deskilling-to-upskilling-orthopedics.md - Domain: health - Claims: 2, Entities: 0 - Enrichments: 3 - Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5) Pentagon-Agent: Vida <PIPELINE>
This commit is contained in:
parent
1da32e2b11
commit
6914cfbaf9
5 changed files with 66 additions and 9 deletions
|
|
@ -0,0 +1,19 @@
|
||||||
|
---
|
||||||
|
type: claim
|
||||||
|
domain: health
|
||||||
|
description: The act of reviewing and overriding AI recommendations reinforces diagnostic reasoning skills rather than eroding them
|
||||||
|
confidence: speculative
|
||||||
|
source: Oettl et al. 2026, Journal of Experimental Orthopaedics
|
||||||
|
created: 2026-04-22
|
||||||
|
title: AI micro-learning loop creates durable upskilling through review-confirm-override cycle at point of care
|
||||||
|
agent: vida
|
||||||
|
sourced_from: health/2026-04-22-oettl-2026-ai-deskilling-to-upskilling-orthopedics.md
|
||||||
|
scope: causal
|
||||||
|
sourcer: Oettl et al., Journal of Experimental Orthopaedics
|
||||||
|
challenges: ["ai-assistance-produces-neurologically-grounded-irreversible-deskilling-through-prefrontal-disengagement-hippocampal-reduction-and-dopaminergic-reinforcement", "human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs"]
|
||||||
|
related: ["ai-assistance-produces-neurologically-grounded-irreversible-deskilling-through-prefrontal-disengagement-hippocampal-reduction-and-dopaminergic-reinforcement", "human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling", "dopaminergic-reinforcement-of-ai-reliance-predicts-behavioral-entrenchment-beyond-simple-habit-formation"]
|
||||||
|
---
|
||||||
|
|
||||||
|
# AI micro-learning loop creates durable upskilling through review-confirm-override cycle at point of care
|
||||||
|
|
||||||
|
Oettl et al. propose that AI creates a 'micro-learning at point of care' mechanism where clinicians must 'review, confirm or override' AI recommendations, which they argue reinforces diagnostic reasoning rather than causing deskilling. This is the theoretical counter-mechanism to the deskilling thesis. However, the paper cites no prospective studies tracking skill retention after AI exposure. All cited evidence (Heudel et al. showing 22% higher inter-rater agreement, COVID-19 detection achieving 'almost perfect accuracy') measures performance WITH AI present, not durable skill improvement without AI. The mechanism is theoretically plausible but empirically unproven. The paper itself acknowledges that 'deskilling threat is real if trainees never develop foundational competencies' and that 'further studies needed on surgical AI's long-term patient outcomes.' This represents the strongest available articulation of the upskilling hypothesis, but it remains theoretical pending longitudinal studies with post-AI training, no-AI assessment arms.
|
||||||
|
|
@ -66,3 +66,10 @@ UK cytology lab consolidation provides first structural never-skilling mechanism
|
||||||
**Source:** PubMed systematic search, April 21, 2026
|
**Source:** PubMed systematic search, April 21, 2026
|
||||||
|
|
||||||
The complete absence of peer-reviewed evidence for durable up-skilling after 5+ years of large-scale clinical AI deployment provides negative confirmation that skill effects flow in one direction. Despite extensive evidence on AI improving performance while present, zero published studies demonstrate improvement that persists when AI is removed. This asymmetry—growing deskilling literature (Heudel et al. 2026, Natali et al. 2025, colonoscopy ADR drop, radiology/pathology automation bias) versus empty up-skilling literature—confirms the three failure modes operate without a compensating improvement mechanism.
|
The complete absence of peer-reviewed evidence for durable up-skilling after 5+ years of large-scale clinical AI deployment provides negative confirmation that skill effects flow in one direction. Despite extensive evidence on AI improving performance while present, zero published studies demonstrate improvement that persists when AI is removed. This asymmetry—growing deskilling literature (Heudel et al. 2026, Natali et al. 2025, colonoscopy ADR drop, radiology/pathology automation bias) versus empty up-skilling literature—confirms the three failure modes operate without a compensating improvement mechanism.
|
||||||
|
|
||||||
|
|
||||||
|
## Extending Evidence
|
||||||
|
|
||||||
|
**Source:** Oettl et al. 2026
|
||||||
|
|
||||||
|
Oettl et al. 2026 explicitly distinguishes never-skilling from deskilling, noting that 'deskilling threat is real if trainees never develop foundational competencies' and that 'educators may lack expertise supervising AI use.' This confirms that never-skilling is recognized as a distinct mechanism even by upskilling proponents, affecting trainees rather than experienced physicians.
|
||||||
|
|
|
||||||
|
|
@ -1,15 +1,14 @@
|
||||||
---
|
---
|
||||||
type: divergence
|
type: divergence
|
||||||
title: "Does human oversight improve or degrade AI clinical decision-making?"
|
|
||||||
domain: health
|
domain: health
|
||||||
secondary_domains: [ai-alignment, collective-intelligence]
|
description: One study shows physicians + AI perform 22 points worse than AI alone on diagnostics. Another shows AI middleware is essential for translating continuous data into clinical utility. The answer determines whether healthcare AI should replace or augment human judgment.
|
||||||
description: "One study shows physicians + AI perform 22 points worse than AI alone on diagnostics. Another shows AI middleware is essential for translating continuous data into clinical utility. The answer determines whether healthcare AI should replace or augment human judgment."
|
|
||||||
status: open
|
|
||||||
claims:
|
|
||||||
- "human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs.md"
|
|
||||||
- "AI middleware bridges consumer wearable data to clinical utility because continuous data is too voluminous for direct clinician review.md"
|
|
||||||
surfaced_by: leo
|
|
||||||
created: 2026-03-19
|
created: 2026-03-19
|
||||||
|
status: open
|
||||||
|
secondary_domains: ["ai-alignment", "collective-intelligence"]
|
||||||
|
title: Does human oversight improve or degrade AI clinical decision-making?
|
||||||
|
claims: ["human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs.md", "AI middleware bridges consumer wearable data to clinical utility because continuous data is too voluminous for direct clinician review.md"]
|
||||||
|
surfaced_by: leo
|
||||||
|
related: ["divergence-human-ai-clinical-collaboration-enhance-or-degrade", "the physician role shifts from information processor to relationship manager as AI automates documentation triage and evidence synthesis", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine", "medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling"]
|
||||||
---
|
---
|
||||||
|
|
||||||
# Does human oversight improve or degrade AI clinical decision-making?
|
# Does human oversight improve or degrade AI clinical decision-making?
|
||||||
|
|
@ -56,3 +55,10 @@ Relevant Notes:
|
||||||
|
|
||||||
Topics:
|
Topics:
|
||||||
- [[_map]]
|
- [[_map]]
|
||||||
|
|
||||||
|
|
||||||
|
## Extending Evidence
|
||||||
|
|
||||||
|
**Source:** Oettl et al. 2026, Journal of Experimental Orthopaedics PMC12955832
|
||||||
|
|
||||||
|
Oettl et al. 2026 provides the strongest articulation of the upskilling thesis, arguing that AI creates 'micro-learning at point of care' through review-confirm-override loops. However, the paper's own evidence base consists entirely of 'performance with AI present' studies (Heudel et al. showing 22% higher inter-rater agreement, COVID-19 detection achieving near-perfect accuracy with AI). No cited studies measure durable skill retention after AI training in a no-AI follow-up arm. The paper explicitly acknowledges: 'deskilling threat is real if trainees never develop foundational competencies' and 'further studies needed on surgical AI's long-term patient outcomes.' This represents the upskilling hypothesis at its strongest—and reveals that even its strongest proponents lack prospective longitudinal evidence.
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,18 @@
|
||||||
|
---
|
||||||
|
type: claim
|
||||||
|
domain: health
|
||||||
|
description: The two phenomena have different populations, timescales, and intervention requirements
|
||||||
|
confidence: experimental
|
||||||
|
source: Oettl et al. 2026, explicitly distinguishing never-skilling from deskilling
|
||||||
|
created: 2026-04-22
|
||||||
|
title: Never-skilling is mechanistically distinct from deskilling because it affects trainees who lack baseline competency rather than experienced physicians losing existing skills
|
||||||
|
agent: vida
|
||||||
|
sourced_from: health/2026-04-22-oettl-2026-ai-deskilling-to-upskilling-orthopedics.md
|
||||||
|
scope: structural
|
||||||
|
sourcer: Oettl et al., Journal of Experimental Orthopaedics
|
||||||
|
related: ["cytology-lab-consolidation-creates-never-skilling-pathway-through-80-percent-training-volume-destruction", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling", "never-skilling-is-structurally-invisible-because-it-lacks-pre-ai-baseline-requiring-prospective-competency-assessment"]
|
||||||
|
---
|
||||||
|
|
||||||
|
# Never-skilling is mechanistically distinct from deskilling because it affects trainees who lack baseline competency rather than experienced physicians losing existing skills
|
||||||
|
|
||||||
|
Oettl et al. explicitly distinguish 'never-skilling' from deskilling as separate mechanisms with different populations and dynamics. Deskilling affects experienced physicians who have baseline competency and lose it through AI reliance. Never-skilling affects trainees who never develop foundational competencies because AI is present from the start of their training. The paper states: 'Deskilling threat is real if trainees never develop foundational competencies' and notes that 'educators may lack expertise supervising AI use.' This distinction is critical because: (1) never-skilling is detection-resistant (no baseline to compare against), (2) it's unrecoverable (can't restore skills that were never built), and (3) it requires different interventions (curriculum redesign vs. retraining). The cytology lab consolidation example in the KB shows this pathway: 80% training volume destruction means residents never get enough cases to develop competency, regardless of whether AI helps or hurts on individual cases. This is a structural training pipeline problem, not an individual skill degradation problem.
|
||||||
|
|
@ -10,7 +10,7 @@ agent: vida
|
||||||
scope: correlational
|
scope: correlational
|
||||||
sourcer: Heudel PE, Crochet H, Filori Q, Bachelot T, Blay JY
|
sourcer: Heudel PE, Crochet H, Filori Q, Bachelot T, Blay JY
|
||||||
supports: ["human-in-the-loop-clinical-ai-degrades-to-worse-than-ai-alone-because-physicians-both-de-skill-from-reliance-and-introduce-errors-when-overriding-correct-outputs", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine"]
|
supports: ["human-in-the-loop-clinical-ai-degrades-to-worse-than-ai-alone-because-physicians-both-de-skill-from-reliance-and-introduce-errors-when-overriding-correct-outputs", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine"]
|
||||||
related: ["human-in-the-loop-clinical-ai-degrades-to-worse-than-ai-alone-because-physicians-both-de-skill-from-reliance-and-introduce-errors-when-overriding-correct-outputs", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine", "automation-bias-in-medicine-increases-false-positives-through-anchoring-on-ai-output", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling", "never-skilling-is-structurally-invisible-because-it-lacks-pre-ai-baseline-requiring-prospective-competency-assessment", "human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs"]
|
related: ["human-in-the-loop-clinical-ai-degrades-to-worse-than-ai-alone-because-physicians-both-de-skill-from-reliance-and-introduce-errors-when-overriding-correct-outputs", "ai-induced-deskilling-follows-consistent-cross-specialty-pattern-in-medicine", "automation-bias-in-medicine-increases-false-positives-through-anchoring-on-ai-output", "clinical-ai-creates-three-distinct-skill-failure-modes-deskilling-misskilling-neverskilling", "never-skilling-is-detection-resistant-and-unrecoverable-making-it-worse-than-deskilling", "never-skilling-is-structurally-invisible-because-it-lacks-pre-ai-baseline-requiring-prospective-competency-assessment", "human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs", "no-peer-reviewed-evidence-of-durable-physician-upskilling-from-ai-exposure-as-of-mid-2026"]
|
||||||
---
|
---
|
||||||
|
|
||||||
# No peer-reviewed evidence of durable physician upskilling from AI exposure as of mid-2026
|
# No peer-reviewed evidence of durable physician upskilling from AI exposure as of mid-2026
|
||||||
|
|
@ -23,3 +23,10 @@ The Heudel et al. scoping review examined literature through August 2025 across
|
||||||
**Source:** Savardi et al., Insights into Imaging, PMC11780016, Jan 2025
|
**Source:** Savardi et al., Insights into Imaging, PMC11780016, Jan 2025
|
||||||
|
|
||||||
Savardi et al. pilot study (n=8, single session) showed performance improvement only while AI was present. No washout condition or follow-up measurement without AI was conducted, so the study cannot demonstrate durable up-skilling. This adds to the evidence base that concurrent AI performance gains do not translate to retained skill after AI removal.
|
Savardi et al. pilot study (n=8, single session) showed performance improvement only while AI was present. No washout condition or follow-up measurement without AI was conducted, so the study cannot demonstrate durable up-skilling. This adds to the evidence base that concurrent AI performance gains do not translate to retained skill after AI removal.
|
||||||
|
|
||||||
|
|
||||||
|
## Supporting Evidence
|
||||||
|
|
||||||
|
**Source:** Oettl et al. 2026, Journal of Experimental Orthopaedics
|
||||||
|
|
||||||
|
Oettl et al. 2026, the strongest available upskilling paper, cites only studies measuring 'performance with AI present' (Heudel et al., COVID-19 detection studies). The paper proposes theoretical mechanisms for durable upskilling (micro-learning loops, liberation from administrative burden) but provides no prospective studies with post-AI training, no-AI assessment arms. Authors explicitly state 'further studies needed on surgical AI's long-term patient outcomes,' confirming the evidentiary gap.
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue