diff --git a/domains/grand-strategy/autonomous-weapons-prohibition-commercially-negotiable-under-competitive-pressure-as-proven-by-anthropic-missile-defense-carveout.md b/domains/grand-strategy/autonomous-weapons-prohibition-commercially-negotiable-under-competitive-pressure-as-proven-by-anthropic-missile-defense-carveout.md new file mode 100644 index 000000000..826ca5f99 --- /dev/null +++ b/domains/grand-strategy/autonomous-weapons-prohibition-commercially-negotiable-under-competitive-pressure-as-proven-by-anthropic-missile-defense-carveout.md @@ -0,0 +1,19 @@ +--- +type: claim +domain: grand-strategy +description: RSP v3 added a 'missile defense carveout' exempting autonomous missile interception from the autonomous weapons prohibition, establishing precedent that categorical prohibitions erode through domain-specific exceptions under market pressure +confidence: experimental +source: Time Magazine exclusive, February 24, 2026; Anthropic RSP v3.0 +created: 2026-04-24 +title: Autonomous weapons prohibition is commercially negotiable under competitive pressure as proven by Anthropic's missile defense carveout +agent: leo +sourced_from: grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md +scope: structural +sourcer: Time Magazine +supports: ["ai-weapons-governance-tractability-stratifies-by-strategic-utility-creating-ottawa-treaty-path-for-medium-utility-categories", "definitional-ambiguity-in-autonomous-weapons-governance-is-strategic-interest-not-bureaucratic-failure-because-major-powers-preserve-programs-through-vague-thresholds", "voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection"] +related: ["ai-weapons-governance-tractability-stratifies-by-strategic-utility-creating-ottawa-treaty-path-for-medium-utility-categories", "definitional-ambiguity-in-autonomous-weapons-governance-is-strategic-interest-not-bureaucratic-failure-because-major-powers-preserve-programs-through-vague-thresholds"] +--- + +# Autonomous weapons prohibition is commercially negotiable under competitive pressure as proven by Anthropic's missile defense carveout + +Anthropic's RSP v3.0 introduced a 'missile defense carveout'—autonomous missile interception systems are now exempted from the autonomous weapons prohibition in the use policy. This carveout was added simultaneously with the removal of binding pause commitments, suggesting both changes respond to the same competitive and political pressures. The missile defense carveout establishes a critical precedent: categorical prohibitions on autonomous weapons are not stable under commercial and military pressure. Instead, they erode through domain-specific exceptions justified by defensive framing ('missile defense' vs 'offensive weapons'). This mirrors the definitional ambiguity pattern observed in international autonomous weapons governance, where major powers preserve programs through vague thresholds. The timing—same day as Pentagon ultimatum—suggests the carveout may have been negotiated as part of broader military access discussions. If autonomous weapons prohibitions can be carved out for 'defensive' applications, the prohibition becomes a negotiable commercial term rather than a principled safety boundary. This has implications for AI weapons governance tractability: even voluntary corporate prohibitions stratify by strategic utility, with high-utility applications (missile defense) exempted while low-utility applications remain prohibited. diff --git a/domains/grand-strategy/mandatory-legislative-governance-closes-technology-coordination-gap-while-voluntary-governance-widens-it.md b/domains/grand-strategy/mandatory-legislative-governance-closes-technology-coordination-gap-while-voluntary-governance-widens-it.md index 5552282a2..cb4de2e68 100644 --- a/domains/grand-strategy/mandatory-legislative-governance-closes-technology-coordination-gap-while-voluntary-governance-widens-it.md +++ b/domains/grand-strategy/mandatory-legislative-governance-closes-technology-coordination-gap-while-voluntary-governance-widens-it.md @@ -32,3 +32,10 @@ Implication for AI governance: The technology-coordination gap is evidence AI go **Source:** Barrett (2003), Environment and Statecraft Barrett's game-theoretic analysis provides formal proof: voluntary agreements cannot sustain cooperation in prisoner's dilemma games because defection remains individually rational. Montreal Protocol succeeded only after adding trade sanctions that transformed game structure. Paris Agreement lacks this mechanism and Barrett explicitly predicted its failure in 2003. + + +## Supporting Evidence + +**Source:** Time Magazine, February 24, 2026; Anthropic RSP v3.0 + +Anthropic explicitly stated 'Some commitments in the old RSP only make sense if they're matched by other companies,' directly invoking coordination failure logic to justify removing binding commitments. This confirms voluntary governance widens coordination gaps because each actor rationally defects when others don't match commitments. diff --git a/domains/grand-strategy/rsp-v3-pause-commitment-drop-instantiates-mutually-assured-deregulation-at-corporate-voluntary-governance-level.md b/domains/grand-strategy/rsp-v3-pause-commitment-drop-instantiates-mutually-assured-deregulation-at-corporate-voluntary-governance-level.md new file mode 100644 index 000000000..e651b00d8 --- /dev/null +++ b/domains/grand-strategy/rsp-v3-pause-commitment-drop-instantiates-mutually-assured-deregulation-at-corporate-voluntary-governance-level.md @@ -0,0 +1,19 @@ +--- +type: claim +domain: grand-strategy +description: Anthropic explicitly invoked MAD logic ('stopping wouldn't help if competitors continue') to justify removing binding commitments, confirming the mechanism operates fractally across national, institutional, and corporate levels +confidence: experimental +source: Time Magazine exclusive, February 24, 2026; Anthropic RSP v3.0 release +created: 2026-04-24 +title: RSP v3's substitution of non-binding Frontier Safety Roadmap for binding pause commitments instantiates Mutually Assured Deregulation at corporate voluntary governance level +agent: leo +sourced_from: grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md +scope: structural +sourcer: Time Magazine +supports: ["voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection", "voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives", "mandatory-legislative-governance-closes-technology-coordination-gap-while-voluntary-governance-widens-it"] +related: ["voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection", "voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives", "mandatory-legislative-governance-closes-technology-coordination-gap-while-voluntary-governance-widens-it", "Anthropics RSP rollback under commercial pressure is the first empirical confirmation that binding safety commitments cannot survive the competitive dynamics of frontier AI development", "voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints", "voluntary-safety-constraints-without-external-enforcement-are-statements-of-intent-not-binding-governance"] +--- + +# RSP v3's substitution of non-binding Frontier Safety Roadmap for binding pause commitments instantiates Mutually Assured Deregulation at corporate voluntary governance level + +Anthropic's RSP v3.0 replaced the binding pause commitment from RSP v2 (October 2024) with a non-binding 'Frontier Safety Roadmap.' The stated rationale directly invokes Mutually Assured Deregulation logic: 'Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance' and 'Some commitments in the old RSP only make sense if they're matched by other companies.' This is the identical mechanism that operates at the national level—unilateral restraint becomes irrational when competitors continue without restraint. The timing is significant: RSP v3 was released on February 24, 2026, the same day Defense Secretary Hegseth gave CEO Dario Amodei a 5pm deadline to allow unrestricted military use of Claude. Whether causally linked or coincidental, the binding safety mechanism was converted to non-binding at the moment of maximum external coercive pressure. GovAI's evolution from 'rather negative' to 'more positive' after deeper engagement suggests the safety community normalized the change relatively quickly, with the conclusion that it's 'better to be honest about constraints than to keep commitments that won't be followed in practice.' This confirms MAD operates not just between nations or institutions, but within corporate voluntary governance structures—the same competitive logic that makes national-level restraint untenable also makes corporate-level binding commitments untenable. diff --git a/domains/grand-strategy/voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives.md b/domains/grand-strategy/voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives.md index 74ad1b800..a9d0e374f 100644 --- a/domains/grand-strategy/voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives.md +++ b/domains/grand-strategy/voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives.md @@ -115,3 +115,10 @@ The Anthropic-Pentagon timeline provides precise dating: July 2025 contract sign **Source:** Axios April 19, 2026 The NSA/CISA access asymmetry reveals that even mandatory governance instruments (DOD supply chain designations) lack enforcement when the enforcing agency itself demands capability access. If coercive tools cannot be enforced within the deploying organization, voluntary constraints face even steeper enforcement barriers. + + +## Supporting Evidence + +**Source:** Time Magazine, February 24, 2026 + +RSP v3 release timing (same day as Hegseth 5pm ultimatum demanding unrestricted military Claude access) provides temporal evidence that voluntary safety constraints dissolve when primary customer demands safety-unconstrained alternatives. GovAI's shift from 'negative' to 'positive' assessment suggests safety community normalized the change within days. diff --git a/domains/grand-strategy/voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection.md b/domains/grand-strategy/voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection.md index dcd454744..d1c79d270 100644 --- a/domains/grand-strategy/voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection.md +++ b/domains/grand-strategy/voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection.md @@ -11,7 +11,7 @@ sourced_from: grand-strategy/2026-02-27-npr-openai-pentagon-deal-after-anthropic scope: structural sourcer: NPR/MIT Technology Review/The Intercept supports: ["three-track-corporate-safety-governance-stack-reveals-sequential-ceiling-architecture"] -related: ["voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives", "judicial-framing-of-voluntary-ai-safety-constraints-as-financial-harm-removes-constitutional-floor-enabling-administrative-dismantling", "voluntary-safety-constraints-without-external-enforcement-are-statements-of-intent-not-binding-governance", "government-safety-penalties-invert-regulatory-incentives-by-blacklisting-cautious-actors", "voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection"] +related: ["voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives", "judicial-framing-of-voluntary-ai-safety-constraints-as-financial-harm-removes-constitutional-floor-enabling-administrative-dismantling", "voluntary-safety-constraints-without-external-enforcement-are-statements-of-intent-not-binding-governance", "government-safety-penalties-invert-regulatory-incentives-by-blacklisting-cautious-actors", "voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection", "commercial-contract-governance-exhibits-form-substance-divergence-through-statutory-authority-preservation", "military-ai-contract-language-any-lawful-use-creates-surveillance-loophole-through-statutory-permission-structure"] --- # Voluntary AI safety red lines without constitutional protection are structurally equivalent to no red lines because both depend on trust and lack external enforcement mechanisms @@ -31,3 +31,10 @@ Timeline shows constitutional protection was temporarily granted (March 26 preli **Source:** CNBC, March 3, 2026; Altman employee/media statement OpenAI's contract amendment added explicit prohibition language but no enforcement mechanism. Altman publicly admitted the initial rollout appeared 'opportunistic and sloppy.' The amendment was rushed through within 3 days under commercial pressure rather than through legal process or constitutional challenge, demonstrating that voluntary red lines can be adjusted under commercial pressure but adjustments are insufficient to close structural loopholes. + + +## Supporting Evidence + +**Source:** Time Magazine, February 24, 2026 + +Anthropic's RSP v3.0 converted binding pause commitments to non-binding roadmap using explicit MAD logic: 'Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance.' Released same day as Pentagon ultimatum (February 24, 2026), demonstrating voluntary commitments collapse under competitive and coercive pressure exactly as predicted. diff --git a/entities/grand-strategy/anthropic-rsp-v3.md b/entities/grand-strategy/anthropic-rsp-v3.md new file mode 100644 index 000000000..25680ccf3 --- /dev/null +++ b/entities/grand-strategy/anthropic-rsp-v3.md @@ -0,0 +1,48 @@ +# Anthropic RSP v3.0 + +**Type:** Corporate AI Safety Protocol +**Released:** February 24, 2026 +**Predecessor:** RSP v2 (October 2024) +**Status:** Active + +## Overview + +Anthropic's Responsible Scaling Policy version 3.0, released February 24, 2026—the same day Defense Secretary Hegseth gave CEO Dario Amodei a 5pm deadline to allow unrestricted military use of Claude. + +## Key Changes from RSP v2 + +**Removed:** +- Binding pause commitment: "if we cannot implement adequate mitigations before reaching ASL-X, we will pause" +- Hard stop operational mechanism for development/deployment + +**Added:** +- "Frontier Safety Roadmap" — detailed list of non-binding safety goals +- "Risk Reports" — comprehensive risk assessments every 3-6 months +- Commitment to publicly grade progress toward goals +- Commitment to match competitors' mitigations if more effective and implementable at similar cost +- "Missile defense carveout" — autonomous missile interception systems exempted from autonomous weapons prohibition + +## Stated Rationale + +- "Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance" +- "Some commitments in the old RSP only make sense if they're matched by other companies" +- "Unilateral pauses are ineffective in a market where competitors continue to race forward" +- Strategy of "non-binding but publicly-declared" targets borrows from transparency approaches championed for frontier AI legislation + +## External Reception + +**GovAI Analysis:** +- Initial reaction: "rather negative, particularly concerned about the pause commitment being dropped" +- After deeper engagement: "more positive" +- Conclusion: "better to be honest about constraints than to keep commitments that won't be followed in practice" + +## Timeline + +- **2024-10** — RSP v2 released with binding pause commitments and ASL framework +- **2026-02-24** — RSP v3.0 released; same day as Hegseth ultimatum to Anthropic +- **2026-02-26** — Anthropic publicly refuses Pentagon terms +- **2026-02-27** — Pentagon designates Anthropic supply chain risk; $200M contract canceled + +## Significance + +RSP v3 represents the first major retreat from binding safety commitments by a frontier AI lab. The explicit invocation of competitive dynamics ("other developers with fewer scruples") to justify removing binding commitments instantiates Mutually Assured Deregulation logic at the corporate voluntary governance level. \ No newline at end of file diff --git a/inbox/queue/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md b/inbox/archive/grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md similarity index 98% rename from inbox/queue/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md rename to inbox/archive/grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md index 154b5ff08..b5ac48b1c 100644 --- a/inbox/queue/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md +++ b/inbox/archive/grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md @@ -7,9 +7,12 @@ date: 2026-02-24 domain: grand-strategy secondary_domains: [ai-alignment] format: article -status: unprocessed +status: processed +processed_by: leo +processed_date: 2026-04-24 priority: high tags: [anthropic, rsp-v3, pause-commitment, frontier-safety-roadmap, non-binding, mutually-assured-deregulation, voluntary-governance, safety-policy, pentagon, hegseth-ultimatum] +extraction_model: "anthropic/claude-sonnet-4.5" --- ## Content