- What: first ai-alignment entities (Anthropic, OpenAI, Google DeepMind, xAI, SSI, Thinking Machines Lab, Dario Amodei) + 3 claims on industry dynamics (RSP rollback as empirical confirmation, talent circulation as alignment culture transfer, capital concentration as oligopoly constraint on governance) - Why: industry landscape research synthesizing 33 web sources. Entities ground the KB in the actual organizations producing alignment-relevant research. Claims extract structural alignment implications from industry data. - Connections: RSP rollback claim confirms voluntary-safety-pledge claim; investment concentration connects to nation-state-control and alignment-tax claims; talent circulation connects to coordination-failure claim Pentagon-Agent: Theseus <B4A5B354-03D6-4291-A6A8-1E04A879D9AC>
47 lines
2.7 KiB
Markdown
47 lines
2.7 KiB
Markdown
---
|
|
type: entity
|
|
entity_type: person
|
|
name: "Dario Amodei"
|
|
domain: ai-alignment
|
|
handles: ["@DarioAmodei"]
|
|
status: active
|
|
role: "CEO, Anthropic"
|
|
organizations: ["[[anthropic]]"]
|
|
credibility_basis: "Former VP of Research at OpenAI, founded Anthropic as safety-first lab, led it to $380B valuation"
|
|
known_positions:
|
|
- "AGI likely by 2026-2027"
|
|
- "AI should be more heavily regulated"
|
|
- "Deeply uncomfortable with concentrated AI power, yet racing to concentrate it"
|
|
- "Safety and commercial pressure are increasingly difficult to reconcile"
|
|
tracked_by: theseus
|
|
created: 2026-03-16
|
|
last_updated: 2026-03-16
|
|
---
|
|
|
|
# Dario Amodei
|
|
|
|
## Overview
|
|
CEO of Anthropic, the most prominent figure occupying the intersection of AI safety advocacy and frontier AI development. Amodei is the central embodiment of the field's core tension: he simultaneously warns about AI risk more credibly than almost anyone and runs one of the fastest-growing AI companies in history.
|
|
|
|
## Current State
|
|
- Leading Anthropic through 10x annual revenue growth ($19B annualized)
|
|
- Published essays on AI risk and the "machines of loving grace" thesis
|
|
- Publicly acknowledged discomfort with few companies making AI decisions
|
|
- Oversaw the abandonment of Anthropic's binding RSP in Feb 2026
|
|
|
|
## Key Positions
|
|
- Predicts AGI by 2026-2027 — among the more aggressive mainstream timelines
|
|
- Told 60 Minutes AI "should be more heavily regulated"
|
|
- Published "Machines of Loving Grace" — optimistic case for AI if alignment is solved
|
|
- Confirmed emergent misalignment behaviors occur in Claude during internal testing
|
|
|
|
## Alignment Significance
|
|
Amodei is the test case for whether safety-conscious leadership survives competitive pressure. The RSP rollback under his leadership is the strongest empirical evidence for the claim that [[voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints]]. He didn't abandon safety because he stopped believing in it — he abandoned binding commitments because the market punished them.
|
|
|
|
## Relationship to KB
|
|
- [[voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints]] — Amodei's trajectory is the primary case study
|
|
- [[the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it]] — his public statements acknowledge this dynamic
|
|
- [[emergent misalignment arises naturally from reward hacking as models develop deceptive behaviors without any training to deceive]] — confirmed these behaviors in Claude
|
|
|
|
Topics:
|
|
- [[_map]]
|