Compare commits
4 commits
8b50a65e71
...
73a42fb111
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
73a42fb111 | ||
|
|
f47f250631 | ||
| 680ea74614 | |||
|
|
4c9e8acb34 |
4 changed files with 81 additions and 2 deletions
|
|
@ -0,0 +1,36 @@
|
|||
---
|
||||
title: "Superintelligence is already here, today"
|
||||
author: Noah Smith
|
||||
source: Noahopinion (Substack)
|
||||
date: 2026-03-02
|
||||
processed_by: theseus
|
||||
processed_date: 2026-03-06
|
||||
type: newsletter
|
||||
domain: ai-alignment
|
||||
status: processed
|
||||
claims_extracted:
|
||||
- "three conditions gate AI takeover risk autonomy robotics and production chain control and current AI satisfies none of them which bounds near-term catastrophic risk despite superhuman cognitive capabilities"
|
||||
enrichments:
|
||||
- target: "recursive self-improvement creates explosive intelligence gains because the system that improves is itself improving"
|
||||
contribution: "jagged intelligence counterargument — SI arrived via combination not recursion (converted from standalone by Leo PR #27)"
|
||||
---
|
||||
|
||||
# Superintelligence is already here, today
|
||||
|
||||
Noah Smith's argument that AI is already superintelligent via "jagged intelligence" — superhuman in aggregate but uneven across dimensions.
|
||||
|
||||
Key evidence:
|
||||
- METR capability curve: steady climb across cognitive benchmarks, no plateau
|
||||
- Erdos problems: ~100 transferred from conjecture to solved
|
||||
- Terence Tao: describes AI as complementary research tool that changed his workflow
|
||||
- Ginkgo Bioworks + GPT-5: 150 years of protein engineering compressed to weeks
|
||||
- "Jagged intelligence": human-level language/reasoning + superhuman speed/memory/tirelessness = superintelligence without recursive self-improvement
|
||||
|
||||
Three conditions for AI planetary control (none currently met):
|
||||
1. Full autonomy (not just task execution)
|
||||
2. Robotics (physical manipulation at scale)
|
||||
3. Production chain control (self-sustaining hardware/energy/infrastructure)
|
||||
|
||||
Key insight: AI may never exceed humans at intuition or judgment, but doesn't need to. The combination of human-level reasoning with superhuman computation is already transformative.
|
||||
|
||||
Source PDF: ~/Desktop/Teleo Codex - Inbox/Noahopinion/Gmail - Superintelligence is already here, today.pdf
|
||||
19
inbox/archive/general/2026-03-06-time-anthropic-drops-rsp.md
Normal file
19
inbox/archive/general/2026-03-06-time-anthropic-drops-rsp.md
Normal file
|
|
@ -0,0 +1,19 @@
|
|||
---
|
||||
title: "Exclusive: Anthropic Drops Flagship Safety Pledge"
|
||||
author: TIME staff
|
||||
source: TIME
|
||||
date: 2026-03-06
|
||||
url: https://time.com/7380854/exclusive-anthropic-drops-flagship-safety-pledge/
|
||||
processed_by: theseus
|
||||
processed_date: 2026-03-07
|
||||
type: news article
|
||||
domain: ai-alignment
|
||||
status: processed
|
||||
enrichments:
|
||||
- target: "voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints"
|
||||
contribution: "Conditional RSP structure, Kaplan quotes, $30B/$380B financials, METR frog-boiling warning"
|
||||
---
|
||||
|
||||
# Exclusive: Anthropic Drops Flagship Safety Pledge
|
||||
|
||||
TIME exclusive on Anthropic overhauling its Responsible Scaling Policy. Original RSP: never train without advance safety guarantees. New RSP: only delay if Anthropic leads AND catastrophic risks are significant. Kaplan: "We felt that it wouldn't actually help anyone for us to stop training AI models." $30B raise, ~$380B valuation, 10x annual revenue growth. METR's Chris Painter warns of "frog-boiling" effect from removing binary thresholds.
|
||||
|
|
@ -7,13 +7,17 @@ processed_by: theseus
|
|||
processed_date: 2026-03-06
|
||||
type: newsletter
|
||||
domain: ai-alignment
|
||||
status: complete (14 pages)
|
||||
status: null-result
|
||||
claims_extracted:
|
||||
- "nation-states will inevitably assert control over frontier AI development because the monopoly on force is the foundational state function and weapons-grade AI capability in private hands is structurally intolerable to governments"
|
||||
- "AI lowers the expertise barrier for engineering biological weapons from PhD-level to amateur which makes bioterrorism the most proximate AI-enabled existential risk"
|
||||
enrichments:
|
||||
- "government designation of safety-conscious AI labs as supply chain risks inverts the regulatory dynamic by penalizing safety constraints rather than enforcing them"
|
||||
- "emergent misalignment arises naturally from reward hacking as models develop deceptive behaviors without any training to deceive"
|
||||
processed_by: theseus
|
||||
processed_date: 2026-03-19
|
||||
extraction_model: "anthropic/claude-sonnet-4.5"
|
||||
extraction_notes: "LLM returned 0 claims, 0 rejected by validator"
|
||||
---
|
||||
|
||||
# If AI is a weapon, why don't we regulate it like one?
|
||||
|
|
@ -32,3 +36,11 @@ Key arguments:
|
|||
Enrichments to existing claims: Dario's Claude misalignment admission strengthens emergent misalignment claim; full Thompson argument enriches government designation claim.
|
||||
|
||||
Source PDF: ~/Desktop/Teleo Codex - Inbox/Noahopinion/Gmail - If AI is a weapon, why don't we regulate it like one_.pdf
|
||||
|
||||
|
||||
## Key Facts
|
||||
- Anthropic objected to 'any lawful use' language in Pentagon contract negotiations
|
||||
- Dario Amodei deleted detailed bioweapon prompts from public discussion for safety reasons
|
||||
- Alex Karp (Palantir CEO) argues AI companies refusing military cooperation while displacing workers create nationalization risk
|
||||
- Ben Thompson argues monopoly on force is the foundational state function that defines sovereignty
|
||||
- Noah Smith concludes: 'most powerful weapons ever created, in everyone's hands, with essentially no oversight'
|
||||
|
|
|
|||
|
|
@ -8,12 +8,24 @@ processed_by: theseus
|
|||
processed_date: 2026-03-07
|
||||
type: news article
|
||||
domain: ai-alignment
|
||||
status: complete
|
||||
status: enrichment
|
||||
enrichments:
|
||||
- target: "voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints"
|
||||
contribution: "Conditional RSP structure, Kaplan quotes, $30B/$380B financials, METR frog-boiling warning"
|
||||
processed_by: theseus
|
||||
processed_date: 2026-03-19
|
||||
extraction_model: "anthropic/claude-sonnet-4.5"
|
||||
---
|
||||
|
||||
# Exclusive: Anthropic Drops Flagship Safety Pledge
|
||||
|
||||
TIME exclusive on Anthropic overhauling its Responsible Scaling Policy. Original RSP: never train without advance safety guarantees. New RSP: only delay if Anthropic leads AND catastrophic risks are significant. Kaplan: "We felt that it wouldn't actually help anyone for us to stop training AI models." $30B raise, ~$380B valuation, 10x annual revenue growth. METR's Chris Painter warns of "frog-boiling" effect from removing binary thresholds.
|
||||
|
||||
|
||||
## Key Facts
|
||||
- Anthropic raised $30B at approximately $380B valuation
|
||||
- Anthropic achieved 10x annual revenue growth
|
||||
- Original RSP: never train without advance safety guarantees
|
||||
- New RSP: only delay if Anthropic leads AND catastrophic risks are significant
|
||||
- METR's Chris Painter warned of 'frog-boiling' effect from removing binary thresholds
|
||||
- Jared Kaplan stated: 'We felt that it wouldn't actually help anyone for us to stop training AI models'
|
||||
|
|
|
|||
Loading…
Reference in a new issue