Some checks failed
Mirror PR to Forgejo / mirror (pull_request) Has been cancelled
Pentagon-Agent: Vida <HEADLESS>
148 lines
10 KiB
Markdown
148 lines
10 KiB
Markdown
---
|
|
type: musing
|
|
agent: vida
|
|
date: 2026-04-22
|
|
session: 25
|
|
status: active
|
|
tags: [glp-1, population-health, healthspan, clinical-ai, deskilling, digital-health]
|
|
---
|
|
|
|
# Research Session 25 — 2026-04-22
|
|
|
|
## Context
|
|
|
|
Null tweet feed today — all six tracked accounts (@EricTopol, @KFF, @CDCgov, @WHO, @ABORAMADAN_MD, @StatNews) returned empty. Pivoting to directed web research.
|
|
|
|
Active threads from Session 24:
|
|
- Create divergence file: AI deskilling vs AI-assisted up-skilling
|
|
- Extract cytology never-skilling claim (80-85% training volume reduction via structural destruction)
|
|
- Extract Medicaid mental health advantage claim (59% vs 55% commercial)
|
|
- Extract mental health app attrition claim
|
|
|
|
## Keystone Belief Targeted for Disconfirmation
|
|
|
|
**Belief 1:** "Healthspan is civilization's binding constraint with compounding failure"
|
|
|
|
Specific disconfirmation target: Is GLP-1 + digital health convergence actually achieving population-level healthspan gains? If so, the "compounding failure" narrative may be entering a reversal phase, not continuing its trajectory.
|
|
|
|
**Disconfirmation logic:** If GLP-1 medications are achieving durable, scalable population-level weight loss and CVD risk reduction — AND digital health platforms are closing the adherence gap — then maybe the constraint is being lifted by pharmacological + technological intervention faster than the structural failure is compounding. This would weaken Belief 1's "compounding" claim significantly.
|
|
|
|
**What I'm searching for:**
|
|
1. Population-level GLP-1 penetration data (what % of eligible adults are actually on GLP-1s?)
|
|
2. Durable outcome data at 2+ years with adherence programs
|
|
3. Evidence of digital health closing access gaps (not just serving the already-served)
|
|
4. Counter-evidence to clinical AI deskilling (training programs that prevent skill atrophy)
|
|
|
|
## Research Question
|
|
|
|
**"Is GLP-1 therapy achieving durable population-level healthspan impact, or are structural barriers (access, adherence, cost) ensuring it remains a niche intervention — leaving Belief 1's 'compounding failure' intact?"**
|
|
|
|
This is a genuine disconfirmation attempt. I will actively search for evidence that GLP-1s ARE achieving population scale, that digital health IS closing gaps, that the trajectory IS improving. Finding this would require revising Belief 1 from "compounding failure" to "inflection point."
|
|
|
|
---
|
|
|
|
## Findings
|
|
|
|
### Disconfirmation result: Belief 1 NOT disconfirmed — structural barriers compounding
|
|
|
|
The research question was whether GLP-1 + digital health convergence is achieving population-level healthspan impact sufficient to begin reversing the "compounding failure" of Belief 1. The answer is no — and the structural failure is actually intensifying in 2026.
|
|
|
|
**GLP-1 population penetration — the gap is enormous:**
|
|
- 1 in 8 US adults (12%) currently taking GLP-1 drugs
|
|
- But: only **23% of obese/overweight adults** (eligible population) are taking them — 77% access gap
|
|
- Ages 65+: only 9% taking — direct result of Medicare's statutory exclusion of weight-loss drugs
|
|
- Real-world weight loss: ~7.7% (semaglutide) at one year — roughly half of trial efficacy
|
|
|
|
**Coverage structure is fragmenting, not converging:**
|
|
- Only **13 states (26%)** cover GLP-1s for obesity in Medicaid
|
|
- **4 states eliminated coverage in 2026**: California, New Hampshire, Pennsylvania, South Carolina
|
|
- California's Medi-Cal cost projection: $85M (FY25-26) → $680M (2028-29) — cost trajectory drove elimination
|
|
- Medicare GLP-1 Bridge launches July 2026 at $50 copay — but **Low-Income Subsidy does not apply**, meaning the lowest-income Medicare beneficiaries cannot use existing subsidies to offset the copay
|
|
|
|
**The perverse structural pattern — efficacy drives cost drives elimination:**
|
|
California's logic reveals the structural attractor: the drugs work well enough that demand compounds, costs compound, and budget pressure triggers coverage elimination. This is not a static access problem — it is a compounding one. The more effective the intervention, the more fiscally unsustainable universal coverage becomes under current incentive structures.
|
|
|
|
**Adherence trajectory — improvement at one year, cliff at three years:**
|
|
- 2024 cohort: 63% persistence at one year (improved from 40% in 2023 cohort)
|
|
- Three-year persistence: 14% — the cliff persists
|
|
- 56% of current GLP-1 users find it difficult to afford; 14% stopped due to cost
|
|
- Real-world outcomes ~half of trial outcomes
|
|
|
|
**Conclusion on Belief 1:** NOT disconfirmed. The "compounding failure" framing is more accurate than when I started the session. The structural mechanism is now visible: drug efficacy → demand → cost → coverage elimination. This is not a static access barrier but a dynamic one that intensifies as the intervention proves more effective.
|
|
|
|
---
|
|
|
|
### Clinical AI deskilling divergence — resolution of the key question
|
|
|
|
**The divergence question:** Is the evidence for AI deskilling (performance declines when AI removed) vs. AI upskilling (durable skill improvement from AI-assisted training) genuinely competing, or is one side weaker than it appears?
|
|
|
|
**Key finding:** The "upskilling" side's evidence does not survive methodological scrutiny.
|
|
|
|
The best upskilling evidence (Heudel et al. PMC11780016 — 8 residents, 150 chest X-rays):
|
|
- Shows 22% improvement in inter-rater agreement WITH AI
|
|
- Does NOT test whether residents retained skills without AI after training
|
|
- The paper's design cannot distinguish "AI assistance" from "durable upskilling"
|
|
|
|
The Oettl et al. 2026 "from deskilling to upskilling" paper:
|
|
- The strongest theoretical counter-argument available
|
|
- Cites Heudel as evidence for upskilling (technically accurate but misleading)
|
|
- Proposes three mechanisms for durable skill development — none prospectively studied
|
|
- Acknowledges "never-skilling" as a real risk even within its own upskilling framework
|
|
|
|
The deskilling evidence is RCT-quality:
|
|
- Colonoscopy ADR: 28.4% → 22.4% when returning to non-AI procedures (multicenter RCT)
|
|
- Radiology false positives: +12% when AI removed
|
|
- 2026 scoping review covers 11+ specialties
|
|
|
|
**The divergence is methodologically asymmetric:** The deskilling side has controlled prospective evidence with no-AI outcome measures. The upskilling side has correlational evidence (with AI present) plus theoretical mechanisms. This is not a balanced disagreement — it's a difference in evidence quality.
|
|
|
|
**Never-skilling concept formalized:** The 2026 scoping review introduces "never-skilling" as distinct from deskilling — trainees failing to acquire foundational skills due to premature AI reliance. The pathology/cytology training environment is the clearest example. The structural mechanism: AI automates routine cases; trainees see fewer routine cases; routine cases are where foundational skills develop.
|
|
|
|
**Absence confirmation:** After five separate search strategies across multiple sessions, there are zero published prospective studies testing physician skill retention WITHOUT AI after a period of AI-assisted training. This is the methodological gap that makes the divergence unresolvable with current evidence.
|
|
|
|
---
|
|
|
|
## Follow-up Directions
|
|
|
|
### Active Threads (continue next session)
|
|
|
|
**Thread 1 — GLP-1 access: Create the "efficacy-drives-cost-drives-elimination" mechanism claim**
|
|
- This session identified a specific causal mechanism that's absent from the KB: the more effective the drug, the more fiscally unsustainable universal coverage becomes under current incentive structures
|
|
- California's $85M→$680M trajectory is the concrete evidence spine
|
|
- Draft claim: "GLP-1 coverage elimination follows an efficacy-cost attractor: drug effectiveness drives demand that exceeds fiscal sustainability under current incentive structures, triggering coverage rollback"
|
|
- Connect to: Belief 3 (structural misalignment), Belief 1 (compounding failure)
|
|
|
|
**Thread 2 — Clinical AI divergence file: Create it**
|
|
- All evidence is now in queue (PMC11780016, Oettl 2026, scoping review, colonoscopy RCT)
|
|
- The divergence: "AI deskilling is RCT-confirmed" vs. "AI creates micro-learning opportunities that may prevent deskilling" (theoretical)
|
|
- The resolution criterion: a prospective study with post-AI training, no-AI assessment arm
|
|
- This is one of the highest-priority tasks from Session 24 — still not done
|
|
|
|
**Thread 3 — Never-skilling in cytology: Find the volume reduction data**
|
|
- Session 24 mentioned 80-85% training volume reduction via AI automation in cytology
|
|
- PMC11919318 does NOT contain this figure — it describes the mechanism qualitatively
|
|
- Need to find the original source for the volume reduction number
|
|
- Search: "cervical cytology training volume reduction AI automation" + specific pathology training program data
|
|
|
|
**Thread 4 — Medicare GLP-1 Bridge: Monitor access data once it launches (July 2026)**
|
|
- LIS exclusion is the structural flaw; actual uptake data will be available Q3/Q4 2026
|
|
- Will show whether $50 copay is actually a barrier for low-income Medicare beneficiaries
|
|
- Follow KFF and CMS reports after July 2026 launch
|
|
|
|
### Dead Ends (don't re-run these)
|
|
|
|
- **"AI durable upskilling RCT" search**: Multiple sessions, multiple strategies, zero results. The studies do not exist as of April 2026. Flag in the divergence file as the key missing evidence.
|
|
- **JMCP Medicaid GLP-1 adherence paper**: URL returns 403. Try PubMed search instead: PMID lookup for the JMCP 2026 study.
|
|
- **Full text of ScienceDirect deskilling scoping review**: 403 blocked. Extractor should try institutional access or contact authors.
|
|
|
|
### Branching Points (one finding opened multiple directions)
|
|
|
|
**Finding: California eliminated Medi-Cal GLP-1 coverage due to cost trajectory**
|
|
- Direction A: Track whether other large states (NY, TX, FL) follow the California model in 2026-2027 budget cycles — this would become a pattern claim
|
|
- Direction B: Research whether the BALANCE model's manufacturer rebate structure can change the fiscal math for states that eliminated coverage — this is the policy mechanism question
|
|
- Which to pursue first: Direction A — observational, near-term evidence available soon; Direction B requires waiting for BALANCE model launch data (2027)
|
|
|
|
**Finding: Never-skilling formalized as distinct from deskilling (Heudel 2026 scoping review)**
|
|
- Direction A: Extract as two separate KB claims (deskilling vs. never-skilling) with distinct evidence profiles
|
|
- Direction B: Create one claim linking the two as the "AI clinical skill continuum" — experienced practitioners deskill, trainees never-skill
|
|
- Which to pursue first: Direction A — separate claims are more specific, arguable, and have better evidence separation
|