vida: extract claims from 2026-04-28-llm-vs-human-glp1-coaching-commoditization-limits
Some checks are pending
Mirror PR to Forgejo / mirror (pull_request) Waiting to run
Some checks are pending
Mirror PR to Forgejo / mirror (pull_request) Waiting to run
- Source: inbox/queue/2026-04-28-llm-vs-human-glp1-coaching-commoditization-limits.md - Domain: health - Claims: 0, Entities: 0 - Enrichments: 4 - Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5) Pentagon-Agent: Vida <PIPELINE>
This commit is contained in:
parent
88702479b7
commit
eaaae598ec
2 changed files with 11 additions and 1 deletions
|
|
@ -58,3 +58,10 @@ Truveta data shows the first 4 weeks (titration phase) are the highest-risk peri
|
|||
**Source:** on/healthcare.tech analysis, Prime Therapeutics via Mercer
|
||||
|
||||
Meta-regression data cited by on/healthcare.tech shows ~50% discontinuation within one year, ~60% weight regain within 12 months of cessation, and 1-in-12 patients (8.3%) remaining on therapy at three years according to Prime Therapeutics data cited by Mercer. This confirms the year-two persistence ceiling and extends the timeline to show continued attrition through year three.
|
||||
|
||||
|
||||
## Extending Evidence
|
||||
|
||||
**Source:** Nicholas Thompson LinkedIn 2026; cross-reference to digital-behavioral-support-improves-glp1-persistence-20-percentage-points
|
||||
|
||||
The $1.8B, 2-person AI-staffed GLP-1 telehealth startup demonstrates that low-end commoditization (prescribing-only, no behavioral support) is already occurring at massive scale. However, this pure-prescribing model likely faces even worse persistence rates than the 14% year-two ceiling, since behavioral support is known to improve GLP-1 persistence by 20 percentage points. The startup's legal issues (FDA warnings, lawsuits over AI-generated patient photos) suggest that AI-only prescribing without behavioral wraparound creates both clinical and legal risks that may limit long-term viability despite short-term revenue growth.
|
||||
|
|
|
|||
|
|
@ -7,11 +7,14 @@ date: 2025-01-01
|
|||
domain: health
|
||||
secondary_domains: [ai-alignment]
|
||||
format: research
|
||||
status: unprocessed
|
||||
status: processed
|
||||
processed_by: vida
|
||||
processed_date: 2026-04-28
|
||||
priority: medium
|
||||
tags: [LLM, AI-coaching, behavioral-support, GLP-1, commoditization, clinical-safety]
|
||||
intake_tier: research-task
|
||||
flagged_for_theseus: ["AI coaching safety: LLM behavioral health applications face same alignment concerns as clinical AI — formulaic responses, bias, privacy — at scale in consumer health context"]
|
||||
extraction_model: "anthropic/claude-sonnet-4.5"
|
||||
---
|
||||
|
||||
## Content
|
||||
Loading…
Reference in a new issue