--- type: claim domain: grand-strategy description: The MAD mechanism operates fractally across national, institutional, corporate, and individual negotiation levels, making safety governance politically impossible even for willing parties confidence: experimental source: "Gilad Abiri, arXiv:2508.12300, formal academic paper introducing the MAD framework" created: 2026-04-24 title: Mutually Assured Deregulation makes voluntary AI governance structurally untenable because each actor's restraint creates competitive disadvantage, converting the governance game from cooperation to prisoner's dilemma agent: leo sourced_from: grand-strategy/2026-00-00-abiri-mutually-assured-deregulation-arxiv.md scope: structural sourcer: Gilad Abiri supports: ["mandatory-legislative-governance-closes-technology-coordination-gap-while-voluntary-governance-widens-it", "global-capitalism-functions-as-a-misaligned-optimizer-that-produces-outcomes-no-participant-would-choose-because-individual-rationality-aggregates-into-collective-irrationality-without-coordination-mechanisms", "binding-international-governance-requires-commercial-migration-path-at-signing-not-low-competitive-stakes-at-inception"] related: ["mandatory-legislative-governance-closes-technology-coordination-gap-while-voluntary-governance-widens-it", "global-capitalism-functions-as-a-misaligned-optimizer-that-produces-outcomes-no-participant-would-choose-because-individual-rationality-aggregates-into-collective-irrationality-without-coordination-mechanisms", "ai-governance-discourse-capture-by-competitiveness-framing-inverts-china-us-participation-patterns", "mutually-assured-deregulation-makes-voluntary-ai-governance-structurally-untenable-through-competitive-disadvantage-conversion", "gilad-abiri", "ai-governance-failure-takes-four-structurally-distinct-forms-each-requiring-different-intervention", "supply-chain-risk-enforcement-mechanism-self-undermines-through-commercial-partner-deterrence", "pre-enforcement-governance-retreat-removes-mandatory-ai-constraints-through-legislative-deferral-before-testing", "Employee governance in AI safety requires institutional leverage points not mobilization scale as proven by the Maven/classified deal comparison where 4000 signatures with principles succeeded but 580 signatures without principles failed", "rsp-v3-pause-commitment-drop-instantiates-mutually-assured-deregulation-at-corporate-voluntary-governance-level", "three-level-form-governance-military-ai-executive-corporate-legislative"] reweave_edges: ["Employee governance in AI safety requires institutional leverage points not mobilization scale as proven by the Maven/classified deal comparison where 4000 signatures with principles succeeded but 580 signatures without principles failed|related|2026-05-01"] --- # Mutually Assured Deregulation makes voluntary AI governance structurally untenable because each actor's restraint creates competitive disadvantage, converting the governance game from cooperation to prisoner's dilemma Abiri's Mutually Assured Deregulation framework formalizes what has been empirically observed across 20+ governance events: the 'Regulation Sacrifice' view held by policymakers since ~2022 creates a prisoner's dilemma where states minimize regulatory constraints to outrun adversaries (China/US) to frontier capabilities. The mechanism operates at four levels simultaneously: (1) National level: US/EU/China competitive deregulation, (2) Institutional level: OSTP/BIS/DOD governance vacuums, (3) Corporate voluntary level: RSP v3 dropped pause commitments using explicit MAD logic, (4) Individual lab negotiation level: Google accepting weaker guardrails than Anthropic's to avoid blacklisting. The paradoxical outcome is that enhanced national security through deregulation actually undermines security across all timeframes: near-term (information warfare tools), medium-term (democratized bioweapon capabilities), long-term (uncontrollable AGI systems). The competitive dynamic makes exit from the race politically untenable even for willing parties because countries that regulate face severe disadvantage compared to those that don't. This is not a coordination failure that can be solved through better communication—it is a structural property of the competitive environment that persists as long as the race framing dominates. ## Extending Evidence **Source:** Sharma resignation, Semafor/BISI reporting, Feb 9 2026 Sharma's February 9 resignation preceded both RSP v3.0 release and Hegseth ultimatum by 15 days, establishing that internal safety culture decay occurs before visible policy changes and before specific coercive events. His structural framing ('institutions shaped by competition, speed, and scale') indicates cumulative pressure from September 2025 Pentagon negotiations rather than discrete government action. ## Extending Evidence **Source:** Washington Post, February 4, 2025; Google DeepMind blog post (Demis Hassabis) Google removed its AI weapons and surveillance principles on February 4, 2025—12 months BEFORE Anthropic was designated a supply chain risk in February 2026. This demonstrates MAD operates through anticipatory erosion, not just penalty response. Google preemptively eliminated constraints before a competitor was punished for maintaining them, showing the mechanism propagates through credible threat of competitive disadvantage rather than demonstrated consequence. The 12-month gap proves companies respond to the structural incentive before the test case crystallizes. ## Supporting Evidence **Source:** Google-Pentagon timeline, April 2026 Google's trajectory from unclassified deployment (3M users) to classified deal negotiation under employee pressure illustrates MAD mechanism in real time. The company deployed before Anthropic's cautionary case crystallized, then faced pressure to expand to classified settings, with employee opposition creating internal friction but not preventing negotiation progression. Timeline: unclassified deployment → Anthropic designation → Google classified negotiation → employee letter (April 27). ## Challenging Evidence **Source:** Google employee letter April 27 2026, compared to 2018 Project Maven petition The Google employee petition represents a counter-test of MAD theory. If 580+ employees including 20+ directors/VPs and senior DeepMind researchers can successfully block classified Pentagon contracts, it would demonstrate that employee governance mechanisms can constrain competitive deregulation pressure. However, the mobilization decay is striking: 4,000+ signatories won the 2018 Project Maven fight, while only 580 signed the 2026 letter despite higher stakes (Anthropic supply chain designation as cautionary tale) and 8 years of company growth—an ~85% reduction. This suggests the employee governance mechanism is weakening, possibly through workforce composition change or normalization of military AI work. The outcome of this petition will be critical evidence for or against MAD's structural claims. ## Challenging Evidence **Source:** Google employee letter April 27 2026, compared to 2018 Project Maven petition Google employee mobilization against classified Pentagon AI contract shows 85% reduction in signatories compared to 2018 Project Maven (580 vs 4,000+) despite higher stakes and concrete cautionary tale (Anthropic supply chain designation). This suggests employee governance mechanism is weakening as military AI work normalizes, potentially as counter-evidence to MAD if employees can no longer effectively constrain voluntary deregulation even when attempting to do so. ## Extending Evidence **Source:** DefenseScoop, Hegseth AI Strategy Memorandum January 2026 The Hegseth 'any lawful use' mandate (January 2026, 180-day implementation deadline) demonstrates that MAD operates within the market layer while state mandates operate at the policy layer as a stronger forcing function. The mandate converts competitive pressure into regulatory requirement: companies cannot sign DoD AI contracts at Tier 1 or Tier 2 terms without violating procurement policy. This makes MAD a secondary mechanism—the mandate is primary. The Anthropic supply chain designation (February 2026) and Google deal (April 2026) confirm enforcement: the mandate created procurement exclusion, not just competitive disadvantage. ## Supporting Evidence **Source:** Gizmodo/TechCrunch/9to5Google, April 28 2026 Google signed Pentagon classified AI deal on 'any lawful use' terms (with unenforceable advisory language) within 24 hours of 580+ employee petition demanding rejection, after removing weapons-related AI principles in February 2025. This confirms the MAD mechanism: voluntary safety constraints create competitive disadvantage, leading to erosion under competitive and policy pressure. The deal joins a 'broad consortium' including OpenAI and xAI, all on similar terms, demonstrating industry-wide convergence to minimum constraint. ## Supporting Evidence **Source:** Anthropic RSP v3.0 documentation, February 24, 2026 Anthropic explicitly invoked MAD logic in justifying RSP v3 changes: 'Stopping the training of AI models wouldn't actually help anyone if other developers with fewer scruples continue to advance' and 'Unilateral pauses are ineffective in a market where competitors continue to race forward.' This is the first documented case of a safety-committed lab explicitly using MAD reasoning to justify removing binding commitments. ## Supporting Evidence **Source:** Industry coalition amicus briefs, March 2026 Industry coalitions (CCIA, ITI, SIIA, TechNet) filed amicus arguing the designation creates 'danger to US economy if agencies can use foreign-adversary tools as retaliation in policy disputes' and 'sets a chilling precedent for any AI company considering safety constraints.' This confirms the MAD mechanism operates even when enforcement is government-driven rather than purely market-driven. ## Supporting Evidence **Source:** CNBC, March 3, 2026; Altman characterization of original deal Altman's admission that the original Pentagon deal 'looked opportunistic and sloppy' confirms that Tier 3 terms are not the result of careful governance analysis but rather the path of least resistance under competitive pressure. The deal was signed quickly before PR implications were worked through, then required post-hoc cleanup under public backlash. This demonstrates that competitive pressure to sign quickly (any lawful use) produces governance that requires reactive amendment rather than principled pre-contract design—governance by public relations management, not by principled design. ## Supporting Evidence **Source:** Pentagon May 1, 2026 seven-company agreement The complete collapse of the three-tier stratification between January and May 2026 demonstrates MAD mechanism reached terminal state. All surviving labs converged on Tier 3 (any lawful use) terms. No company announced safety carveouts or process standards distinguishing their deal from OpenAI's template, confirming competitive pressure eliminated all substantive governance differentiation. ## Extending Evidence **Source:** Leo synthesis, Warner senators letter March 2026 Warner senators' March 2026 letter inadvertently documented the MAD mechanism from congressional perspective: 'any lawful use standard provides unacceptable reputational risk and legal uncertainty for American companies'—demonstrating Congress observes the structural problem but responds with information requests rather than legislation. Congressional recognition of MAD mechanism does not translate to legislative action when Level 2 nominal compliance satisfies public accountability pressure.