Commit graph

2 commits

Author SHA1 Message Date
39d7bf5f98 theseus: extract from 3 Dario/Anthropic sources — 3 enrichments + 2 claims
- What: 3 enrichments to existing claims + 2 new standalone claims + 3 source archives
- Sources: TIME "Anthropic Drops Flagship Safety Pledge" (Mar 2026),
  Dario Amodei "Machines of Loving Grace" (darioamodei.com),
  Dario Amodei "The Adolescence of Technology" (darioamodei.com)

Enrichments:
1. voluntary safety pledges claim: Conditional RSP structure (only pause if
   leading AND catastrophic), Kaplan quotes, $30B/$380B financials, METR
   frog-boiling warning
2. bioterrorism claim: Anthropic mid-2025 measurements (2-3x uplift),
   STEM-degree threshold approaching, 36/38 gene synthesis providers fail
   screening, mirror life extinction scenario, ASL-3 classification
3. RSI claim: AI already writing much of Anthropic's code, 1-2 years from
   current gen autonomously building next gen

New claims:
1. AI personas from pre-training as spectrum of humanlike motivations —
   challenges monomaniacal goal models (experimental)
2. Marginal returns to intelligence bounded by five complementary factors —
   bounds what SI can achieve (likely)

Cross-domain flags: health (compressed 21st century), internet-finance
(labor displacement, GDP growth), foundations (chip export controls,
civilizational maturation)

Source diversity note: 3 sources from Dario Amodei / Anthropic — correlated
priors flagged per >3 rule

Pentagon-Agent: Theseus <845F10FB-BC22-40F6-A6A6-F6E4D8F78465>
2026-03-06 15:02:34 +00:00
235d12d0a2 theseus: add 3 claims from Anthropic/Pentagon/nuclear news + enrich 2 foundations
New claims:
- voluntary safety pledges collapse under competitive pressure (Anthropic RSP rollback Feb 2026)
- government supply chain designation penalizes safety (Pentagon/Anthropic Mar 2026)
- models escalate to nuclear war 95% of the time (King's College war games Feb 2026)

Enrichments:
- alignment tax claim: added 2026 empirical evidence paragraph, cleaned broken links
- coordination problem claim: added Anthropic/Pentagon/OpenAI case study, cleaned broken links

Pentagon-Agent: Theseus <845F10FB-BC22-40F6-A6A6-F6E4D8F78465>

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-03-06 12:41:42 +00:00