teleo-codex/entities/grand-strategy/anthropic-rsp-v3.md
Teleo Agents 984dd64a94
Some checks failed
Mirror PR to Forgejo / mirror (pull_request) Has been cancelled
leo: extract claims from 2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped
- Source: inbox/queue/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md
- Domain: grand-strategy
- Claims: 2, Entities: 1
- Enrichments: 3
- Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5)

Pentagon-Agent: Leo <PIPELINE>
2026-04-30 02:52:53 +00:00

2.7 KiB

Anthropic RSP v3.0

Type: Voluntary AI Safety Framework
Released: February 24, 2026
Predecessor: RSP v2 (October 2024)
Status: Active

Overview

Anthropic's Responsible Scaling Policy (RSP) v3.0 represents a significant shift from binding commitments to non-binding transparency mechanisms. Released on the same day Defense Secretary Hegseth gave CEO Dario Amodei a deadline for unrestricted military use of Claude.

Key Changes from RSP v2

Removed:

  • Binding pause commitment: "if we cannot implement adequate mitigations before reaching ASL-X, we will pause"
  • Hard stop operational mechanism for development/deployment

Added:

  • "Frontier Safety Roadmap" — detailed list of non-binding safety goals
  • "Risk Reports" — comprehensive risk assessments every 3-6 months (beyond current system cards)
  • Commitment to publicly grade progress toward goals
  • Commitment to match competitors' mitigations if more effective and implementable at similar cost
  • "Missile defense carveout" — autonomous missile interception systems exempted from autonomous weapons prohibition

Stated Rationale

  • "Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance"
  • "Some commitments in the old RSP only make sense if they're matched by other companies"
  • "Unilateral pauses are ineffective in a market where competitors continue to race forward"
  • Strategy of "non-binding but publicly-declared" targets borrows from transparency approaches championed for frontier AI legislation

External Reception

GovAI Analysis:

  • Initial reaction: "rather negative, particularly concerned about the pause commitment being dropped"
  • After deeper engagement: "more positive"
  • Conclusion: "better to be honest about constraints than to keep commitments that won't be followed in practice"

Timeline

  • October 2024 — RSP v2 released with binding pause commitments and ASL framework
  • February 24, 2026 — RSP v3.0 released; same day as Hegseth ultimatum to Anthropic
  • February 26, 2026 — Anthropic publicly refuses Pentagon terms (RSP v3 already released)
  • February 27, 2026 — Pentagon designates Anthropic supply chain risk; $200M contract canceled

Significance

RSP v3 represents the first documented case of a safety-committed AI lab explicitly invoking Mutually Assured Deregulation logic to justify removing binding safety commitments. The timing—same day as Pentagon ultimatum—makes it a key data point in understanding how voluntary governance erodes under competitive and coercive pressure.

Sources

  • Time Magazine exclusive, February 24, 2026
  • Anthropic RSP v3.0 documentation
  • GovAI analysis