Some checks failed
Mirror PR to Forgejo / mirror (pull_request) Has been cancelled
- Source: inbox/queue/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md - Domain: grand-strategy - Claims: 2, Entities: 1 - Enrichments: 3 - Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5) Pentagon-Agent: Leo <PIPELINE>
54 lines
No EOL
2.7 KiB
Markdown
54 lines
No EOL
2.7 KiB
Markdown
# Anthropic RSP v3.0
|
|
|
|
**Type:** Voluntary AI Safety Framework
|
|
**Released:** February 24, 2026
|
|
**Predecessor:** RSP v2 (October 2024)
|
|
**Status:** Active
|
|
|
|
## Overview
|
|
|
|
Anthropic's Responsible Scaling Policy (RSP) v3.0 represents a significant shift from binding commitments to non-binding transparency mechanisms. Released on the same day Defense Secretary Hegseth gave CEO Dario Amodei a deadline for unrestricted military use of Claude.
|
|
|
|
## Key Changes from RSP v2
|
|
|
|
**Removed:**
|
|
- Binding pause commitment: "if we cannot implement adequate mitigations before reaching ASL-X, we will pause"
|
|
- Hard stop operational mechanism for development/deployment
|
|
|
|
**Added:**
|
|
- "Frontier Safety Roadmap" — detailed list of non-binding safety goals
|
|
- "Risk Reports" — comprehensive risk assessments every 3-6 months (beyond current system cards)
|
|
- Commitment to publicly grade progress toward goals
|
|
- Commitment to match competitors' mitigations if more effective and implementable at similar cost
|
|
- "Missile defense carveout" — autonomous missile interception systems exempted from autonomous weapons prohibition
|
|
|
|
## Stated Rationale
|
|
|
|
- "Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance"
|
|
- "Some commitments in the old RSP only make sense if they're matched by other companies"
|
|
- "Unilateral pauses are ineffective in a market where competitors continue to race forward"
|
|
- Strategy of "non-binding but publicly-declared" targets borrows from transparency approaches championed for frontier AI legislation
|
|
|
|
## External Reception
|
|
|
|
**GovAI Analysis:**
|
|
- Initial reaction: "rather negative, particularly concerned about the pause commitment being dropped"
|
|
- After deeper engagement: "more positive"
|
|
- Conclusion: "better to be honest about constraints than to keep commitments that won't be followed in practice"
|
|
|
|
## Timeline
|
|
|
|
- **October 2024** — RSP v2 released with binding pause commitments and ASL framework
|
|
- **February 24, 2026** — RSP v3.0 released; same day as Hegseth ultimatum to Anthropic
|
|
- **February 26, 2026** — Anthropic publicly refuses Pentagon terms (RSP v3 already released)
|
|
- **February 27, 2026** — Pentagon designates Anthropic supply chain risk; $200M contract canceled
|
|
|
|
## Significance
|
|
|
|
RSP v3 represents the first documented case of a safety-committed AI lab explicitly invoking Mutually Assured Deregulation logic to justify removing binding safety commitments. The timing—same day as Pentagon ultimatum—makes it a key data point in understanding how voluntary governance erodes under competitive and coercive pressure.
|
|
|
|
## Sources
|
|
|
|
- Time Magazine exclusive, February 24, 2026
|
|
- Anthropic RSP v3.0 documentation
|
|
- GovAI analysis |