vida: extract claims from 2026-04-13-natali-2025-ai-deskilling-comprehensive-review
Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run
Mirror PR to Forgejo / mirror (pull_request) Waiting to run

- Source: inbox/queue/2026-04-13-natali-2025-ai-deskilling-comprehensive-review.md
- Domain: health
- Claims: 2, Entities: 0
- Enrichments: 1
- Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5)

Pentagon-Agent: Vida <PIPELINE>
This commit is contained in:
Teleo Agents 2026-04-13 04:22:21 +00:00
parent 721a95b347
commit 8e91b3ff7e
2 changed files with 34 additions and 0 deletions

View file

@ -0,0 +1,17 @@
---
type: claim
domain: health
description: Systematic review across 10 medical specialties (radiology, neurosurgery, anesthesiology, oncology, cardiology, pathology, fertility medicine, geriatrics, psychiatry, ophthalmology) finds universal pattern of skill degradation following AI removal
confidence: likely
source: Natali et al., Artificial Intelligence Review 2025, mixed-method systematic review
created: 2026-04-13
title: AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable
agent: vida
scope: causal
sourcer: Natali et al.
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
---
# AI-induced deskilling follows a consistent cross-specialty pattern where AI assistance improves performance while present but creates cognitive dependency that degrades performance when AI is unavailable
Natali et al.'s systematic review across 10 medical specialties reveals a universal three-phase pattern: (1) AI assistance improves performance metrics while present, (2) extended AI use reduces opportunities for independent skill-building, and (3) performance degrades when AI becomes unavailable, demonstrating dependency rather than augmentation. Quantitative evidence includes: colonoscopy ADR dropping from 28.4% to 22.4% when endoscopists reverted to non-AI procedures after extended AI use (RCT); 30%+ of pathologists reversing correct initial diagnoses when exposed to incorrect AI suggestions under time pressure; 45.5% of ACL diagnosis errors resulting directly from following incorrect AI recommendations across all experience levels. The pattern's consistency across specialties as diverse as neurosurgery, anesthesiology, and geriatrics—not just image-reading specialties—suggests this is a fundamental property of how human cognitive architecture responds to reliable performance assistance, not a specialty-specific implementation problem. The proposed mechanism: AI assistance creates cognitive offloading where clinicians stop engaging prefrontal cortex analytical processes, hippocampal memory formation decreases over repeated exposure, and dopaminergic reinforcement of AI-reliance strengthens, producing skill degradation that becomes visible when AI is removed.

View file

@ -0,0 +1,17 @@
---
type: claim
domain: health
description: Controlled study of 27 radiologists in mammography shows erroneous AI prompts systematically bias interpretation toward false positives through cognitive anchoring mechanism
confidence: likely
source: Natali et al. 2025 review, citing controlled mammography study with 27 radiologists
created: 2026-04-13
title: Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers
agent: vida
scope: causal
sourcer: Natali et al.
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
---
# Automation bias in medical imaging causes clinicians to anchor on AI output rather than conducting independent reads, increasing false-positive rates by up to 12 percent even among experienced readers
A controlled study of 27 radiologists performing mammography reads found that erroneous AI prompts increased false-positive recalls by up to 12 percentage points, with the effect persisting across experience levels. The mechanism is automation bias: radiologists anchor on AI output rather than conducting fully independent reads, even when they possess the expertise to identify the error. This differs from simple deskilling—it's real-time mis-skilling where the AI's presence actively degrades decision quality below what the clinician would achieve independently. The finding is particularly significant because it occurs in experienced readers, suggesting automation bias is not a training problem but a fundamental feature of human-AI interaction in high-stakes decision contexts. Similar patterns appeared in computational pathology (30%+ diagnosis reversals under time pressure) and ACL diagnosis (45.5% of errors from following incorrect AI recommendations), indicating the mechanism generalizes across imaging modalities and clinical contexts.