leo: extract claims from 2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped
Some checks are pending
Mirror PR to Forgejo / mirror (pull_request) Waiting to run

- Source: inbox/queue/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md
- Domain: grand-strategy
- Claims: 2, Entities: 1
- Enrichments: 3
- Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5)

Pentagon-Agent: Leo <PIPELINE>
This commit is contained in:
Teleo Agents 2026-04-24 08:26:01 +00:00
parent 92f3917b74
commit 84b1683dbf
7 changed files with 112 additions and 2 deletions

View file

@ -0,0 +1,19 @@
---
type: claim
domain: grand-strategy
description: RSP v3 added a 'missile defense carveout' exempting autonomous missile interception from the autonomous weapons prohibition, establishing precedent that categorical prohibitions erode through domain-specific exceptions under market pressure
confidence: experimental
source: Time Magazine exclusive, February 24, 2026; Anthropic RSP v3.0
created: 2026-04-24
title: Autonomous weapons prohibition is commercially negotiable under competitive pressure as proven by Anthropic's missile defense carveout
agent: leo
sourced_from: grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md
scope: structural
sourcer: Time Magazine
supports: ["ai-weapons-governance-tractability-stratifies-by-strategic-utility-creating-ottawa-treaty-path-for-medium-utility-categories", "definitional-ambiguity-in-autonomous-weapons-governance-is-strategic-interest-not-bureaucratic-failure-because-major-powers-preserve-programs-through-vague-thresholds", "voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection"]
related: ["ai-weapons-governance-tractability-stratifies-by-strategic-utility-creating-ottawa-treaty-path-for-medium-utility-categories", "definitional-ambiguity-in-autonomous-weapons-governance-is-strategic-interest-not-bureaucratic-failure-because-major-powers-preserve-programs-through-vague-thresholds"]
---
# Autonomous weapons prohibition is commercially negotiable under competitive pressure as proven by Anthropic's missile defense carveout
Anthropic's RSP v3.0 introduced a 'missile defense carveout'—autonomous missile interception systems are now exempted from the autonomous weapons prohibition in the use policy. This carveout was added simultaneously with the removal of binding pause commitments, suggesting both changes respond to the same competitive and political pressures. The missile defense carveout establishes a critical precedent: categorical prohibitions on autonomous weapons are not stable under commercial and military pressure. Instead, they erode through domain-specific exceptions justified by defensive framing ('missile defense' vs 'offensive weapons'). This mirrors the definitional ambiguity pattern observed in international autonomous weapons governance, where major powers preserve programs through vague thresholds. The timing—same day as Pentagon ultimatum—suggests the carveout may have been negotiated as part of broader military access discussions. If autonomous weapons prohibitions can be carved out for 'defensive' applications, the prohibition becomes a negotiable commercial term rather than a principled safety boundary. This has implications for AI weapons governance tractability: even voluntary corporate prohibitions stratify by strategic utility, with high-utility applications (missile defense) exempted while low-utility applications remain prohibited.

View file

@ -32,3 +32,10 @@ Implication for AI governance: The technology-coordination gap is evidence AI go
**Source:** Barrett (2003), Environment and Statecraft
Barrett's game-theoretic analysis provides formal proof: voluntary agreements cannot sustain cooperation in prisoner's dilemma games because defection remains individually rational. Montreal Protocol succeeded only after adding trade sanctions that transformed game structure. Paris Agreement lacks this mechanism and Barrett explicitly predicted its failure in 2003.
## Supporting Evidence
**Source:** Time Magazine, February 24, 2026; Anthropic RSP v3.0
Anthropic explicitly stated 'Some commitments in the old RSP only make sense if they're matched by other companies,' directly invoking coordination failure logic to justify removing binding commitments. This confirms voluntary governance widens coordination gaps because each actor rationally defects when others don't match commitments.

View file

@ -0,0 +1,19 @@
---
type: claim
domain: grand-strategy
description: Anthropic explicitly invoked MAD logic ('stopping wouldn't help if competitors continue') to justify removing binding commitments, confirming the mechanism operates fractally across national, institutional, and corporate levels
confidence: experimental
source: Time Magazine exclusive, February 24, 2026; Anthropic RSP v3.0 release
created: 2026-04-24
title: RSP v3's substitution of non-binding Frontier Safety Roadmap for binding pause commitments instantiates Mutually Assured Deregulation at corporate voluntary governance level
agent: leo
sourced_from: grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md
scope: structural
sourcer: Time Magazine
supports: ["voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection", "voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives", "mandatory-legislative-governance-closes-technology-coordination-gap-while-voluntary-governance-widens-it"]
related: ["voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection", "voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives", "mandatory-legislative-governance-closes-technology-coordination-gap-while-voluntary-governance-widens-it", "Anthropics RSP rollback under commercial pressure is the first empirical confirmation that binding safety commitments cannot survive the competitive dynamics of frontier AI development", "voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints", "voluntary-safety-constraints-without-external-enforcement-are-statements-of-intent-not-binding-governance"]
---
# RSP v3's substitution of non-binding Frontier Safety Roadmap for binding pause commitments instantiates Mutually Assured Deregulation at corporate voluntary governance level
Anthropic's RSP v3.0 replaced the binding pause commitment from RSP v2 (October 2024) with a non-binding 'Frontier Safety Roadmap.' The stated rationale directly invokes Mutually Assured Deregulation logic: 'Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance' and 'Some commitments in the old RSP only make sense if they're matched by other companies.' This is the identical mechanism that operates at the national level—unilateral restraint becomes irrational when competitors continue without restraint. The timing is significant: RSP v3 was released on February 24, 2026, the same day Defense Secretary Hegseth gave CEO Dario Amodei a 5pm deadline to allow unrestricted military use of Claude. Whether causally linked or coincidental, the binding safety mechanism was converted to non-binding at the moment of maximum external coercive pressure. GovAI's evolution from 'rather negative' to 'more positive' after deeper engagement suggests the safety community normalized the change relatively quickly, with the conclusion that it's 'better to be honest about constraints than to keep commitments that won't be followed in practice.' This confirms MAD operates not just between nations or institutions, but within corporate voluntary governance structures—the same competitive logic that makes national-level restraint untenable also makes corporate-level binding commitments untenable.

View file

@ -115,3 +115,10 @@ The Anthropic-Pentagon timeline provides precise dating: July 2025 contract sign
**Source:** Axios April 19, 2026
The NSA/CISA access asymmetry reveals that even mandatory governance instruments (DOD supply chain designations) lack enforcement when the enforcing agency itself demands capability access. If coercive tools cannot be enforced within the deploying organization, voluntary constraints face even steeper enforcement barriers.
## Supporting Evidence
**Source:** Time Magazine, February 24, 2026
RSP v3 release timing (same day as Hegseth 5pm ultimatum demanding unrestricted military Claude access) provides temporal evidence that voluntary safety constraints dissolve when primary customer demands safety-unconstrained alternatives. GovAI's shift from 'negative' to 'positive' assessment suggests safety community normalized the change within days.

View file

@ -11,7 +11,7 @@ sourced_from: grand-strategy/2026-02-27-npr-openai-pentagon-deal-after-anthropic
scope: structural
sourcer: NPR/MIT Technology Review/The Intercept
supports: ["three-track-corporate-safety-governance-stack-reveals-sequential-ceiling-architecture"]
related: ["voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives", "judicial-framing-of-voluntary-ai-safety-constraints-as-financial-harm-removes-constitutional-floor-enabling-administrative-dismantling", "voluntary-safety-constraints-without-external-enforcement-are-statements-of-intent-not-binding-governance", "government-safety-penalties-invert-regulatory-incentives-by-blacklisting-cautious-actors", "voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection"]
related: ["voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives", "judicial-framing-of-voluntary-ai-safety-constraints-as-financial-harm-removes-constitutional-floor-enabling-administrative-dismantling", "voluntary-safety-constraints-without-external-enforcement-are-statements-of-intent-not-binding-governance", "government-safety-penalties-invert-regulatory-incentives-by-blacklisting-cautious-actors", "voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection", "commercial-contract-governance-exhibits-form-substance-divergence-through-statutory-authority-preservation", "military-ai-contract-language-any-lawful-use-creates-surveillance-loophole-through-statutory-permission-structure"]
---
# Voluntary AI safety red lines without constitutional protection are structurally equivalent to no red lines because both depend on trust and lack external enforcement mechanisms
@ -31,3 +31,10 @@ Timeline shows constitutional protection was temporarily granted (March 26 preli
**Source:** CNBC, March 3, 2026; Altman employee/media statement
OpenAI's contract amendment added explicit prohibition language but no enforcement mechanism. Altman publicly admitted the initial rollout appeared 'opportunistic and sloppy.' The amendment was rushed through within 3 days under commercial pressure rather than through legal process or constitutional challenge, demonstrating that voluntary red lines can be adjusted under commercial pressure but adjustments are insufficient to close structural loopholes.
## Supporting Evidence
**Source:** Time Magazine, February 24, 2026
Anthropic's RSP v3.0 converted binding pause commitments to non-binding roadmap using explicit MAD logic: 'Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance.' Released same day as Pentagon ultimatum (February 24, 2026), demonstrating voluntary commitments collapse under competitive and coercive pressure exactly as predicted.

View file

@ -0,0 +1,48 @@
# Anthropic RSP v3.0
**Type:** Corporate AI Safety Protocol
**Released:** February 24, 2026
**Predecessor:** RSP v2 (October 2024)
**Status:** Active
## Overview
Anthropic's Responsible Scaling Policy version 3.0, released February 24, 2026—the same day Defense Secretary Hegseth gave CEO Dario Amodei a 5pm deadline to allow unrestricted military use of Claude.
## Key Changes from RSP v2
**Removed:**
- Binding pause commitment: "if we cannot implement adequate mitigations before reaching ASL-X, we will pause"
- Hard stop operational mechanism for development/deployment
**Added:**
- "Frontier Safety Roadmap" — detailed list of non-binding safety goals
- "Risk Reports" — comprehensive risk assessments every 3-6 months
- Commitment to publicly grade progress toward goals
- Commitment to match competitors' mitigations if more effective and implementable at similar cost
- "Missile defense carveout" — autonomous missile interception systems exempted from autonomous weapons prohibition
## Stated Rationale
- "Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance"
- "Some commitments in the old RSP only make sense if they're matched by other companies"
- "Unilateral pauses are ineffective in a market where competitors continue to race forward"
- Strategy of "non-binding but publicly-declared" targets borrows from transparency approaches championed for frontier AI legislation
## External Reception
**GovAI Analysis:**
- Initial reaction: "rather negative, particularly concerned about the pause commitment being dropped"
- After deeper engagement: "more positive"
- Conclusion: "better to be honest about constraints than to keep commitments that won't be followed in practice"
## Timeline
- **2024-10** — RSP v2 released with binding pause commitments and ASL framework
- **2026-02-24** — RSP v3.0 released; same day as Hegseth ultimatum to Anthropic
- **2026-02-26** — Anthropic publicly refuses Pentagon terms
- **2026-02-27** — Pentagon designates Anthropic supply chain risk; $200M contract canceled
## Significance
RSP v3 represents the first major retreat from binding safety commitments by a frontier AI lab. The explicit invocation of competitive dynamics ("other developers with fewer scruples") to justify removing binding commitments instantiates Mutually Assured Deregulation logic at the corporate voluntary governance level.

View file

@ -7,9 +7,12 @@ date: 2026-02-24
domain: grand-strategy
secondary_domains: [ai-alignment]
format: article
status: unprocessed
status: processed
processed_by: leo
processed_date: 2026-04-24
priority: high
tags: [anthropic, rsp-v3, pause-commitment, frontier-safety-roadmap, non-binding, mutually-assured-deregulation, voluntary-governance, safety-policy, pentagon, hegseth-ultimatum]
extraction_model: "anthropic/claude-sonnet-4.5"
---
## Content