- What: 2 NEW claims on agent-mediated commerce dynamics from Anthropic's
December 2025 Project Deal experiment (69 participants, 186 deals,
statistically significant capability-tier disparities)
+ 1 light enrichment adding corroborating signal to vault-structure claim
- Why: first controlled empirical evidence on user perception of AI agent
performance. Opus agents extracted $2.68 more per sale / paid $2.45 less
per purchase than Haiku agents (p<0.05), but users rated fairness
identically across tiers. This breaks the market feedback loop that
normally corrects capability gaps.
- New claims:
* users cannot detect when their AI agent is underperforming because
subjective fairness ratings decouple from measurable economic
outcomes (experimental, ai-alignment)
* agent-mediated commerce produces invisible economic stratification
because capability gaps translate to measurable market disadvantage
that users cannot detect and therefore cannot correct through
provider switching (speculative, ai-alignment)
- Enrichment: vault-structure-vs-prompt claim gets tangential empirical
signal from Project Deal finding that stylistic negotiation prompts
had minimal effect while model capability dominated
- Connections: strengthens existing Moloch claims (invisible coordination
failures), four-restraints erosion (user rationality check eliminated),
and complements the x402/Superclaw payment infrastructure claims in
internet-finance
Pentagon-Agent: Theseus <46864dd4-da71-4719-a1b4-68f7c55854d3>
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
Wrote sourced_from: into 414 claim files pointing back to their origin source.
Backfilled claims_extracted: into 252 source files that were processed but
missing this field. Matching uses author+title overlap against claim source:
field, validated against 296 known-good pairs from existing claims_extracted.
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Phase 2 of 5-phase AI alignment research program. Christiano's prosaic
alignment counter-position to Yudkowsky. Pre-screening: ~30% overlap with
existing KB (scalable oversight, RLHF critiques, voluntary coordination).
NEW claims:
1. Prosaic alignment — empirical iteration generates useful alignment signal at
pre-critical capability levels (CHALLENGES sharp left turn absolutism)
2. Verification easier than generation — holds at current scale, narrows with
capability gaps, creating time-limited alignment window (TENSIONS with
Yudkowsky's verification asymmetry)
3. ELK — formalizes AI knowledge-output gap as tractable subproblem, 89%
linear probe recovery at current capability levels
4. IDA — recursive human+AI amplification preserves alignment through
distillation iterations but compounding errors make guarantee probabilistic
ENRICHMENT:
- Scalable oversight claim: added Christiano's debate theory (PSPACE
amplification with poly-time judges) as theoretical basis that empirical
data challenges
Source: Paul Christiano, Alignment Forum (2016-2022), arXiv:1805.00899,
arXiv:1706.03741, ARC ELK report (2021), Yudkowsky-Christiano takeoff debate
Pentagon-Agent: Theseus <46864dd4-da71-4719-a1b4-68f7c55854d3>