teleo-codex/inbox/archive/2026-03-16-theseus-ai-coordination-governance-evidence.md
m3taversal d0998a23bd theseus: AI coordination governance evidence — 3 claims + 1 entity
- What: 3 claims on coordination governance empirics (binding regulation as
  only mechanism that works, transparency declining, compute export controls
  as misaligned governance) + UK AISI entity + comprehensive source archive
- Why: targeted research on weakest grounding of B2 ("alignment is coordination
  problem"). Found that voluntary coordination has empirically failed across
  every mechanism tested (2023-2026). Only binding regulation with enforcement
  changes behavior. This challenges the optimistic version of B2 and
  strengthens the case for enforcement-backed coordination.
- Connections: confirms voluntary-safety-pledge claim with extensive new
  evidence, strengthens nation-state-control claim, challenges alignment-tax
  claim by showing the tax is being cut not paid

Pentagon-Agent: Theseus <B4A5B354-03D6-4291-A6A8-1E04A879D9AC>
2026-03-16 19:35:00 +00:00

3.3 KiB

type title author url date_published date_archived domain status processed_by tags sourced_via
source Empirical Evidence: AI Coordination and Governance Mechanisms That Changed Behavior Theseus research agent (multi-source web synthesis) null 2026-03-16 2026-03-16 ai-alignment processing theseus
ai-governance
coordination
safety-commitments
regulation
enforcement
voluntary-pledges
Theseus research agent — 45 web searches synthesized from Brookings, Stanford FMTI, EU legislation, OECD, government publications, TechCrunch, TIME, CNN, Fortune, academic papers

Empirical Evidence: AI Coordination and Governance Mechanisms That Changed Behavior

Core finding: almost no international AI governance mechanism has produced verified behavioral change at frontier AI labs. Only three mechanisms work: (1) binding regulation with enforcement teeth (EU AI Act, China), (2) export controls backed by state power, (3) competitive/reputational pressure through markets.

Behavioral Change Tier List

Tier 1 — Verified behavioral change:

  • EU AI Act: Apple paused Apple Intelligence in EU, Meta changed ads, EUR 500M+ fines (DMA). Companies preemptively modifying products.
  • China's AI regulations: mandatory algorithm filing, content labeling, criminal enforcement. First binding generative AI regulation (Aug 2023).
  • US export controls: most impactful mechanism. Tiered country system, deployment caps, Nvidia designing compliance chips. Geopolitically motivated, not safety-motivated.

Tier 2 — Institutional infrastructure, uncertain behavioral change:

  • AI Safety Institutes (UK, US, Japan, Korea, Canada). US-UK joint o1 evaluation. But no blocking authority, US AISI defunded/rebranded.
  • Third-party evaluation (METR, Apollo Research). Fragile, no regulatory mandate.

Tier 3 — Partial voluntary compliance:

  • Watermarking: 38% implementation. Google SynthID, Meta AudioSeal. Anthropic the only major lab without one.
  • Red-teaming: self-reported, limited external verification.

Tier 4 — No verified behavioral change:

  • ALL international declarations (Bletchley, Seoul, Paris, Hiroshima, OECD, UN)
  • Frontier Model Forum
  • White House voluntary commitments

Key Evidence Points

  • Stanford FMTI transparency scores DECLINING: -17 points mean (2024→2025). Meta -29, Mistral -37, OpenAI -14.
  • OpenAI explicitly made safety conditional on competitor behavior (Preparedness Framework v2, Apr 2025).
  • OpenAI removed "safely" from mission statement (Nov 2025).
  • OpenAI dissolved Superalignment team (May 2024) and Mission Alignment team (Feb 2026).
  • Google accused by 60 UK lawmakers of violating Seoul commitments (Gemini 2.5 Pro, Apr 2025).
  • 450+ organizations lobbied on AI in 2025 (up from 6 in 2016). $92M in lobbying fees Q1-Q3 2025.
  • SB 1047 (CA AI safety bill) vetoed after heavy industry lobbying.
  • Anthropic's own language: RSP "very hard to meet without industry-wide coordination."

Novel Mechanisms

  • Compute governance: export controls work but geopolitically motivated. KYC for compute proposed, not implemented.
  • Insurance/liability: market projected $29.7B by 2033. Creates market incentives aligned with safety.
  • Third-party auditing: METR, Apollo Research. Apollo warns ecosystem unsustainable without regulatory mandate.
  • Futarchy: implemented for DAO governance (MetaDAO, Optimism experiment) but not yet for AI governance.