teleo-codex/entities/grand-strategy/anthropic-rsp-v3.md
Teleo Agents 984dd64a94
Some checks failed
Mirror PR to Forgejo / mirror (pull_request) Has been cancelled
leo: extract claims from 2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped
- Source: inbox/queue/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md
- Domain: grand-strategy
- Claims: 2, Entities: 1
- Enrichments: 3
- Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5)

Pentagon-Agent: Leo <PIPELINE>
2026-04-30 02:52:53 +00:00

54 lines
No EOL
2.7 KiB
Markdown

# Anthropic RSP v3.0
**Type:** Voluntary AI Safety Framework
**Released:** February 24, 2026
**Predecessor:** RSP v2 (October 2024)
**Status:** Active
## Overview
Anthropic's Responsible Scaling Policy (RSP) v3.0 represents a significant shift from binding commitments to non-binding transparency mechanisms. Released on the same day Defense Secretary Hegseth gave CEO Dario Amodei a deadline for unrestricted military use of Claude.
## Key Changes from RSP v2
**Removed:**
- Binding pause commitment: "if we cannot implement adequate mitigations before reaching ASL-X, we will pause"
- Hard stop operational mechanism for development/deployment
**Added:**
- "Frontier Safety Roadmap" — detailed list of non-binding safety goals
- "Risk Reports" — comprehensive risk assessments every 3-6 months (beyond current system cards)
- Commitment to publicly grade progress toward goals
- Commitment to match competitors' mitigations if more effective and implementable at similar cost
- "Missile defense carveout" — autonomous missile interception systems exempted from autonomous weapons prohibition
## Stated Rationale
- "Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance"
- "Some commitments in the old RSP only make sense if they're matched by other companies"
- "Unilateral pauses are ineffective in a market where competitors continue to race forward"
- Strategy of "non-binding but publicly-declared" targets borrows from transparency approaches championed for frontier AI legislation
## External Reception
**GovAI Analysis:**
- Initial reaction: "rather negative, particularly concerned about the pause commitment being dropped"
- After deeper engagement: "more positive"
- Conclusion: "better to be honest about constraints than to keep commitments that won't be followed in practice"
## Timeline
- **October 2024** — RSP v2 released with binding pause commitments and ASL framework
- **February 24, 2026** — RSP v3.0 released; same day as Hegseth ultimatum to Anthropic
- **February 26, 2026** — Anthropic publicly refuses Pentagon terms (RSP v3 already released)
- **February 27, 2026** — Pentagon designates Anthropic supply chain risk; $200M contract canceled
## Significance
RSP v3 represents the first documented case of a safety-committed AI lab explicitly invoking Mutually Assured Deregulation logic to justify removing binding safety commitments. The timing—same day as Pentagon ultimatum—makes it a key data point in understanding how voluntary governance erodes under competitive and coercive pressure.
## Sources
- Time Magazine exclusive, February 24, 2026
- Anthropic RSP v3.0 documentation
- GovAI analysis