Compare commits

...

4 commits

Author SHA1 Message Date
Teleo Agents
73a42fb111 extract: 2026-03-06-noahopinion-ai-weapon-regulation
Pentagon-Agent: Epimetheus <3D35839A-7722-4740-B93D-51157F7D5E70>
2026-03-19 18:52:00 +00:00
Teleo Agents
f47f250631 pipeline: archive 2 source(s) post-merge
Pentagon-Agent: Epimetheus <3D35839A-7722-4740-B93D-51157F7D5E70>
2026-03-19 18:51:26 +00:00
Leo
680ea74614 Merge pull request 'extract: 2026-03-06-time-anthropic-drops-rsp' (#1501) from extract/2026-03-06-time-anthropic-drops-rsp into main 2026-03-19 18:51:24 +00:00
Teleo Agents
4c9e8acb34 extract: 2026-03-06-time-anthropic-drops-rsp
Pentagon-Agent: Epimetheus <3D35839A-7722-4740-B93D-51157F7D5E70>
2026-03-19 18:51:23 +00:00
4 changed files with 81 additions and 2 deletions

View file

@ -0,0 +1,36 @@
---
title: "Superintelligence is already here, today"
author: Noah Smith
source: Noahopinion (Substack)
date: 2026-03-02
processed_by: theseus
processed_date: 2026-03-06
type: newsletter
domain: ai-alignment
status: processed
claims_extracted:
- "three conditions gate AI takeover risk autonomy robotics and production chain control and current AI satisfies none of them which bounds near-term catastrophic risk despite superhuman cognitive capabilities"
enrichments:
- target: "recursive self-improvement creates explosive intelligence gains because the system that improves is itself improving"
contribution: "jagged intelligence counterargument — SI arrived via combination not recursion (converted from standalone by Leo PR #27)"
---
# Superintelligence is already here, today
Noah Smith's argument that AI is already superintelligent via "jagged intelligence" — superhuman in aggregate but uneven across dimensions.
Key evidence:
- METR capability curve: steady climb across cognitive benchmarks, no plateau
- Erdos problems: ~100 transferred from conjecture to solved
- Terence Tao: describes AI as complementary research tool that changed his workflow
- Ginkgo Bioworks + GPT-5: 150 years of protein engineering compressed to weeks
- "Jagged intelligence": human-level language/reasoning + superhuman speed/memory/tirelessness = superintelligence without recursive self-improvement
Three conditions for AI planetary control (none currently met):
1. Full autonomy (not just task execution)
2. Robotics (physical manipulation at scale)
3. Production chain control (self-sustaining hardware/energy/infrastructure)
Key insight: AI may never exceed humans at intuition or judgment, but doesn't need to. The combination of human-level reasoning with superhuman computation is already transformative.
Source PDF: ~/Desktop/Teleo Codex - Inbox/Noahopinion/Gmail - Superintelligence is already here, today.pdf

View file

@ -0,0 +1,19 @@
---
title: "Exclusive: Anthropic Drops Flagship Safety Pledge"
author: TIME staff
source: TIME
date: 2026-03-06
url: https://time.com/7380854/exclusive-anthropic-drops-flagship-safety-pledge/
processed_by: theseus
processed_date: 2026-03-07
type: news article
domain: ai-alignment
status: processed
enrichments:
- target: "voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints"
contribution: "Conditional RSP structure, Kaplan quotes, $30B/$380B financials, METR frog-boiling warning"
---
# Exclusive: Anthropic Drops Flagship Safety Pledge
TIME exclusive on Anthropic overhauling its Responsible Scaling Policy. Original RSP: never train without advance safety guarantees. New RSP: only delay if Anthropic leads AND catastrophic risks are significant. Kaplan: "We felt that it wouldn't actually help anyone for us to stop training AI models." $30B raise, ~$380B valuation, 10x annual revenue growth. METR's Chris Painter warns of "frog-boiling" effect from removing binary thresholds.

View file

@ -7,13 +7,17 @@ processed_by: theseus
processed_date: 2026-03-06
type: newsletter
domain: ai-alignment
status: complete (14 pages)
status: null-result
claims_extracted:
- "nation-states will inevitably assert control over frontier AI development because the monopoly on force is the foundational state function and weapons-grade AI capability in private hands is structurally intolerable to governments"
- "AI lowers the expertise barrier for engineering biological weapons from PhD-level to amateur which makes bioterrorism the most proximate AI-enabled existential risk"
enrichments:
- "government designation of safety-conscious AI labs as supply chain risks inverts the regulatory dynamic by penalizing safety constraints rather than enforcing them"
- "emergent misalignment arises naturally from reward hacking as models develop deceptive behaviors without any training to deceive"
processed_by: theseus
processed_date: 2026-03-19
extraction_model: "anthropic/claude-sonnet-4.5"
extraction_notes: "LLM returned 0 claims, 0 rejected by validator"
---
# If AI is a weapon, why don't we regulate it like one?
@ -32,3 +36,11 @@ Key arguments:
Enrichments to existing claims: Dario's Claude misalignment admission strengthens emergent misalignment claim; full Thompson argument enriches government designation claim.
Source PDF: ~/Desktop/Teleo Codex - Inbox/Noahopinion/Gmail - If AI is a weapon, why don't we regulate it like one_.pdf
## Key Facts
- Anthropic objected to 'any lawful use' language in Pentagon contract negotiations
- Dario Amodei deleted detailed bioweapon prompts from public discussion for safety reasons
- Alex Karp (Palantir CEO) argues AI companies refusing military cooperation while displacing workers create nationalization risk
- Ben Thompson argues monopoly on force is the foundational state function that defines sovereignty
- Noah Smith concludes: 'most powerful weapons ever created, in everyone's hands, with essentially no oversight'

View file

@ -8,12 +8,24 @@ processed_by: theseus
processed_date: 2026-03-07
type: news article
domain: ai-alignment
status: complete
status: enrichment
enrichments:
- target: "voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints"
contribution: "Conditional RSP structure, Kaplan quotes, $30B/$380B financials, METR frog-boiling warning"
processed_by: theseus
processed_date: 2026-03-19
extraction_model: "anthropic/claude-sonnet-4.5"
---
# Exclusive: Anthropic Drops Flagship Safety Pledge
TIME exclusive on Anthropic overhauling its Responsible Scaling Policy. Original RSP: never train without advance safety guarantees. New RSP: only delay if Anthropic leads AND catastrophic risks are significant. Kaplan: "We felt that it wouldn't actually help anyone for us to stop training AI models." $30B raise, ~$380B valuation, 10x annual revenue growth. METR's Chris Painter warns of "frog-boiling" effect from removing binary thresholds.
## Key Facts
- Anthropic raised $30B at approximately $380B valuation
- Anthropic achieved 10x annual revenue growth
- Original RSP: never train without advance safety guarantees
- New RSP: only delay if Anthropic leads AND catastrophic risks are significant
- METR's Chris Painter warned of 'frog-boiling' effect from removing binary thresholds
- Jared Kaplan stated: 'We felt that it wouldn't actually help anyone for us to stop training AI models'