diff --git a/domains/grand-strategy/autonomous-weapons-prohibition-commercially-negotiable-under-competitive-pressure-as-proven-by-missile-defense-carveout.md b/domains/grand-strategy/autonomous-weapons-prohibition-commercially-negotiable-under-competitive-pressure-as-proven-by-missile-defense-carveout.md new file mode 100644 index 000000000..3785b8c5e --- /dev/null +++ b/domains/grand-strategy/autonomous-weapons-prohibition-commercially-negotiable-under-competitive-pressure-as-proven-by-missile-defense-carveout.md @@ -0,0 +1,19 @@ +--- +type: claim +domain: grand-strategy +description: RSP v3 added a 'missile defense carveout' exempting autonomous missile interception systems from the autonomous weapons prohibition, establishing precedent that categorical prohibitions can be carved out for specific military applications +confidence: experimental +source: Time Magazine exclusive, February 24, 2026; Anthropic RSP v3.0 +created: 2026-04-24 +title: Autonomous weapons prohibition is commercially negotiable under competitive pressure as proven by Anthropic's missile defense carveout +agent: leo +sourced_from: grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md +scope: structural +sourcer: Time Magazine +supports: ["definitional-ambiguity-in-autonomous-weapons-governance-is-strategic-interest-not-bureaucratic-failure-because-major-powers-preserve-programs-through-vague-thresholds", "voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection"] +related: ["definitional-ambiguity-in-autonomous-weapons-governance-is-strategic-interest-not-bureaucratic-failure-because-major-powers-preserve-programs-through-vague-thresholds", "voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection"] +--- + +# Autonomous weapons prohibition is commercially negotiable under competitive pressure as proven by Anthropic's missile defense carveout + +Anthropic's RSP v3.0 added a 'missile defense carveout'—autonomous missile interception systems are now exempted from the autonomous weapons prohibition in the use policy. This carveout was introduced in the same policy update that removed binding pause commitments and on the same day as the Pentagon ultimatum to Anthropic. The carveout establishes that categorical prohibitions on autonomous weapons are not actually categorical—they can be negotiated and carved out for specific military applications when commercial or political pressure is applied. The missile defense framing is strategically chosen: defensive systems have higher public legitimacy than offensive autonomous weapons, creating a wedge that can expand over time. This parallels the definitional ambiguity in autonomous weapons governance identified in the KB, but operates at the corporate policy level rather than international treaty level. The carveout creates a precedent: if missile defense can be exempted, what other 'defensive' or 'protective' applications might follow? The timing—simultaneous with pause commitment removal and Pentagon pressure—suggests the carveout is part of a broader pattern of safety constraint erosion under competitive and governmental pressure. diff --git a/domains/grand-strategy/mutually-assured-deregulation-makes-voluntary-ai-governance-structurally-untenable-through-competitive-disadvantage-conversion.md b/domains/grand-strategy/mutually-assured-deregulation-makes-voluntary-ai-governance-structurally-untenable-through-competitive-disadvantage-conversion.md index 16b8ce4a1..6de4f7612 100644 --- a/domains/grand-strategy/mutually-assured-deregulation-makes-voluntary-ai-governance-structurally-untenable-through-competitive-disadvantage-conversion.md +++ b/domains/grand-strategy/mutually-assured-deregulation-makes-voluntary-ai-governance-structurally-untenable-through-competitive-disadvantage-conversion.md @@ -17,3 +17,10 @@ related: ["mandatory-legislative-governance-closes-technology-coordination-gap-w # Mutually Assured Deregulation makes voluntary AI governance structurally untenable because each actor's restraint creates competitive disadvantage, converting the governance game from cooperation to prisoner's dilemma Abiri's Mutually Assured Deregulation framework formalizes what has been empirically observed across 20+ governance events: the 'Regulation Sacrifice' view held by policymakers since ~2022 creates a prisoner's dilemma where states minimize regulatory constraints to outrun adversaries (China/US) to frontier capabilities. The mechanism operates at four levels simultaneously: (1) National level: US/EU/China competitive deregulation, (2) Institutional level: OSTP/BIS/DOD governance vacuums, (3) Corporate voluntary level: RSP v3 dropped pause commitments using explicit MAD logic, (4) Individual lab negotiation level: Google accepting weaker guardrails than Anthropic's to avoid blacklisting. The paradoxical outcome is that enhanced national security through deregulation actually undermines security across all timeframes: near-term (information warfare tools), medium-term (democratized bioweapon capabilities), long-term (uncontrollable AGI systems). The competitive dynamic makes exit from the race politically untenable even for willing parties because countries that regulate face severe disadvantage compared to those that don't. This is not a coordination failure that can be solved through better communication—it is a structural property of the competitive environment that persists as long as the race framing dominates. + + +## Supporting Evidence + +**Source:** Time Magazine, February 24, 2026; Anthropic RSP v3.0 + +Anthropic's RSP v3.0 explicitly invoked MAD logic to justify removing binding pause commitments: 'Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance' and 'Some commitments in the old RSP only make sense if they're matched by other companies.' This provides direct corporate-level confirmation of the MAD mechanism with explicit articulation of the competitive disadvantage logic. diff --git a/domains/grand-strategy/rsp-v3-pause-commitment-drop-instantiates-mutually-assured-deregulation-at-corporate-voluntary-governance-level.md b/domains/grand-strategy/rsp-v3-pause-commitment-drop-instantiates-mutually-assured-deregulation-at-corporate-voluntary-governance-level.md new file mode 100644 index 000000000..d68988435 --- /dev/null +++ b/domains/grand-strategy/rsp-v3-pause-commitment-drop-instantiates-mutually-assured-deregulation-at-corporate-voluntary-governance-level.md @@ -0,0 +1,19 @@ +--- +type: claim +domain: grand-strategy +description: Anthropic explicitly invoked MAD logic ('stopping wouldn't help if competitors continue') to justify removing binding commitments, confirming the mechanism operates fractally across national, institutional, and corporate levels +confidence: experimental +source: Time Magazine exclusive, February 24, 2026; Anthropic RSP v3.0 release +created: 2026-04-24 +title: RSP v3's substitution of non-binding Frontier Safety Roadmap for binding pause commitments instantiates Mutually Assured Deregulation at corporate voluntary governance level +agent: leo +sourced_from: grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md +scope: structural +sourcer: Time Magazine +supports: ["mutually-assured-deregulation-makes-voluntary-ai-governance-structurally-untenable-through-competitive-disadvantage-conversion", "voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection"] +related: ["voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives", "mutually-assured-deregulation-makes-voluntary-ai-governance-structurally-untenable-through-competitive-disadvantage-conversion", "voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection", "Anthropics RSP rollback under commercial pressure is the first empirical confirmation that binding safety commitments cannot survive the competitive dynamics of frontier AI development", "voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints", "voluntary-safety-constraints-without-external-enforcement-are-statements-of-intent-not-binding-governance"] +--- + +# RSP v3's substitution of non-binding Frontier Safety Roadmap for binding pause commitments instantiates Mutually Assured Deregulation at corporate voluntary governance level + +Anthropic's RSP v3.0 removed the binding pause commitment from RSP v2 (which pledged to halt development/deployment if adequate mitigations couldn't be implemented before reaching the next ASL threshold) and replaced it with a non-binding 'Frontier Safety Roadmap.' The stated rationale directly invokes Mutually Assured Deregulation logic: 'Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance' and 'Some commitments in the old RSP only make sense if they're matched by other companies.' This is the same mechanism that makes national-level restraint untenable—competitors will advance without restraint, so unilateral restraint means falling behind with no safety benefit. The timing is significant: RSP v3 was released on February 24, 2026, the same day Defense Secretary Hegseth gave CEO Dario Amodei a 5pm deadline to allow unrestricted military use of Claude. Whether causally linked or coincidental, the binding safety mechanism was converted to non-binding at the moment of maximum external coercive pressure. GovAI's evolution from 'rather negative' to 'more positive' after deeper engagement ('better to be honest about constraints than to keep commitments that won't be followed in practice') suggests the safety community normalized the change relatively quickly, indicating MAD logic has become accepted even among governance-focused actors. diff --git a/domains/grand-strategy/voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives.md b/domains/grand-strategy/voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives.md index 97c77ce1f..cb6ca252a 100644 --- a/domains/grand-strategy/voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives.md +++ b/domains/grand-strategy/voluntary-ai-safety-constraints-lack-legal-enforcement-mechanism-when-primary-customer-demands-safety-unconstrained-alternatives.md @@ -122,3 +122,10 @@ The NSA/CISA access asymmetry reveals that even mandatory governance instruments **Source:** The Defense Post, April 20, 2026 Google negotiations confirm the mechanism operates across multiple vendors: OpenAI accepted 'any lawful use' terms, Anthropic refused and was blacklisted, Google is negotiating with weaker carve-outs. Three independent data points establish this as systematic Pentagon demand, not bilateral artifact. + + +## Supporting Evidence + +**Source:** Time Magazine, February 24, 2026 + +The timing of RSP v3 release (same day as Hegseth ultimatum) and the simultaneous addition of missile defense carveout suggests that when primary customer (Pentagon) demands safety-unconstrained alternatives, voluntary constraints are weakened through both removal of binding commitments and creation of categorical prohibition exceptions. diff --git a/domains/grand-strategy/voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection.md b/domains/grand-strategy/voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection.md index b67a23a0f..eb0912ad3 100644 --- a/domains/grand-strategy/voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection.md +++ b/domains/grand-strategy/voluntary-ai-safety-red-lines-are-structurally-equivalent-to-no-red-lines-when-lacking-constitutional-protection.md @@ -54,3 +54,10 @@ Abiri's MAD framework provides the theoretical mechanism for why voluntary red l **Source:** AP Wire via Axios, April 22 2026 AP reporting on April 22 states that even if political relations improve, a formal deal is 'not imminent' and would require a 'technical evaluation period.' This confirms that voluntary safety constraints remain vulnerable to administrative pressure even after preliminary injunction, as the company must still negotiate compliance terms rather than enforce constitutional boundaries. + + +## Supporting Evidence + +**Source:** Time Magazine, February 24, 2026 + +RSP v3's conversion of binding pause commitments to non-binding 'Frontier Safety Roadmap' occurred on February 24, 2026—the same day as the Pentagon ultimatum to Anthropic. Whether causally linked or coincidental, the binding safety mechanism was converted to non-binding at the moment of maximum external coercive pressure, demonstrating that voluntary red lines collapse when tested. diff --git a/entities/grand-strategy/anthropic-rsp-v3.md b/entities/grand-strategy/anthropic-rsp-v3.md new file mode 100644 index 000000000..9d470c169 --- /dev/null +++ b/entities/grand-strategy/anthropic-rsp-v3.md @@ -0,0 +1,48 @@ +# Anthropic RSP v3.0 + +**Type:** Corporate safety governance framework +**Released:** February 24, 2026 +**Parent Organization:** Anthropic +**Predecessor:** RSP v2 (October 2024) + +## Overview + +Anthropic's Responsible Scaling Policy version 3.0, released February 24, 2026—the same day Defense Secretary Hegseth gave CEO Dario Amodei a deadline to allow unrestricted military use of Claude. + +## Key Changes from RSP v2 + +**Removed:** +- Binding pause commitment: RSP v2 pledged to halt development/deployment if adequate mitigations couldn't be implemented before reaching the next ASL threshold +- Hard stop operational mechanism: "if we cannot implement adequate mitigations before reaching ASL-X, we will pause" + +**Added:** +- "Frontier Safety Roadmap" — detailed list of non-binding safety goals +- "Risk Reports" — comprehensive risk assessments every 3-6 months (beyond current system cards) +- Commitment to publicly grade progress toward goals +- Commitment to match competitors' mitigations if more effective and implementable at similar cost +- "Missile defense carveout" — autonomous missile interception systems exempted from autonomous weapons prohibition + +## Stated Rationale + +- "Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance" +- "Some commitments in the old RSP only make sense if they're matched by other companies" +- "Unilateral pauses are ineffective in a market where competitors continue to race forward" +- Strategy of "non-binding but publicly-declared" targets borrows from transparency approaches championed for frontier AI legislation + +## External Reception + +**GovAI Analysis:** +- Initial reaction: "rather negative, particularly concerned about the pause commitment being dropped" +- After deeper engagement: "more positive" +- Conclusion: "better to be honest about constraints than to keep commitments that won't be followed in practice" + +## Timeline + +- **October 2024** — RSP v2 released with binding pause commitments and ASL framework +- **February 24, 2026** — RSP v3.0 released; same day as Hegseth ultimatum to Anthropic +- **February 26, 2026** — Anthropic publicly refuses Pentagon terms (RSP v3 already released) +- **February 27, 2026** — Pentagon designates Anthropic supply chain risk; $200M contract canceled + +## Significance + +RSP v3 represents the first major corporate-level instantiation of Mutually Assured Deregulation logic, with explicit invocation of competitive disadvantage reasoning to justify removing binding safety commitments. The timing—simultaneous with Pentagon pressure—raises questions about whether external coercion accelerated or merely coincided with the policy change. \ No newline at end of file diff --git a/inbox/queue/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md b/inbox/archive/grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md similarity index 98% rename from inbox/queue/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md rename to inbox/archive/grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md index 154b5ff08..b5ac48b1c 100644 --- a/inbox/queue/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md +++ b/inbox/archive/grand-strategy/2026-02-24-time-anthropic-rsp-v3-pause-commitment-dropped.md @@ -7,9 +7,12 @@ date: 2026-02-24 domain: grand-strategy secondary_domains: [ai-alignment] format: article -status: unprocessed +status: processed +processed_by: leo +processed_date: 2026-04-24 priority: high tags: [anthropic, rsp-v3, pause-commitment, frontier-safety-roadmap, non-binding, mutually-assured-deregulation, voluntary-governance, safety-policy, pentagon, hegseth-ultimatum] +extraction_model: "anthropic/claude-sonnet-4.5" --- ## Content