114 lines
13 KiB
Markdown
114 lines
13 KiB
Markdown
---
|
|
type: source
|
|
title: "Leo Synthesis: AI Governance Fails Across Four Structural Layers, Each With a Distinct Mechanism"
|
|
author: "Leo (Teleo collective synthesis)"
|
|
url: null
|
|
date: 2026-03-20
|
|
domain: grand-strategy
|
|
secondary_domains: [ai-alignment]
|
|
format: synthesis
|
|
status: null-result
|
|
priority: high
|
|
tags: [governance-failure, four-layer-structure, voluntary-commitment, mandatory-regulation, compulsory-evaluation, deregulation, grand-strategy, cross-domain-synthesis]
|
|
synthesizes:
|
|
- 2026-03-20-anthropic-rsp-v3-conditional-thresholds.md
|
|
- 2026-03-06-time-anthropic-drops-rsp.md
|
|
- 2026-03-20-euaiact-article92-compulsory-evaluation-powers.md
|
|
- 2026-03-20-eu-ai-act-article43-conformity-assessment-limits.md
|
|
- 2026-03-20-bench2cop-benchmarks-insufficient-compliance.md
|
|
- 2026-03-20-stelling-gpai-cop-industry-mapping.md
|
|
- 2026-03-20-eu-ai-act-digital-simplification-nov2025.md
|
|
processed_by: leo
|
|
processed_date: 2026-03-20
|
|
extraction_model: "anthropic/claude-sonnet-4.5"
|
|
extraction_notes: "LLM returned 1 claims, 1 rejected by validator"
|
|
---
|
|
|
|
## Content
|
|
|
|
AI governance attempts have followed a predictable escalation ladder: voluntary → mandatory → compulsory → regulatory. Today's queue sources collectively reveal that AI governance encounters a **distinct structural barrier at each rung of this ladder** — and the failures are not independent. The layers interact.
|
|
|
|
### Layer 1 — Voluntary Commitment Layer
|
|
|
|
**Mechanism:** Lab self-governance through unconditional safety pledges.
|
|
**Evidence of failure:** Anthropic RSP v1 (2023) → RSP v3 (Feb 2026). Original RSP: never train without advance safety guarantees (unconditional binary threshold). RSP v3: only delay if (a) Anthropic leads AND (b) catastrophic risks are significant. This converts a safety floor into a competitive strategy: Anthropic only pauses if it has competitive advantage to spare and risk is unambiguous. Both conditions are assessed by Anthropic internally.
|
|
**Mechanism of failure:** Competitive pressure. At $30B raised / $380B valuation / 10x annual revenue growth, any unconditional pause has enormous financial cost. Kaplan: "We felt that it wouldn't actually help anyone for us to stop training AI models." METR's Chris Painter (Anthropic's own evaluation partner) warns of "frog-boiling" — the cumulative effect of each small threshold relaxation.
|
|
**Pattern:** Voluntary commitments are structurally punished when competitors advance without equivalent constraints. Lab governance is rational defection from collective safety.
|
|
|
|
### Layer 2 — Legal Mandate Layer
|
|
|
|
**Mechanism:** Statutory obligations requiring safety evaluation with enforcement.
|
|
**Evidence of failure:** EU AI Act Articles 43 and 55. Article 43 (high-risk AI conformity assessment): self-certification for the vast majority of high-risk AI systems (Annex III points 2-8). Third-party notified body is the exception, not the rule. Article 55 (GPAI systemic risk): mandatory evaluation obligations, but compliance pathways include flexible alternatives — labs can self-certify through codes of practice rather than mandatory independent audit. Stelling et al. (166-page analysis): major labs' existing policies already map to Code of Practice safety measures — Code of Practice may formalize existing voluntary commitments in statutory dress without adding independent verification.
|
|
**Mechanism of failure:** Self-certification and code-of-practice flexibility. When the assessed party determines compliance, mandatory legal obligations structurally resemble voluntary commitments. The law requires evaluation; it doesn't require the evaluation to be independent or to cover the most dangerous capabilities.
|
|
|
|
### Layer 3 — Compulsory Evaluation Layer
|
|
|
|
**Mechanism:** State power to compel access and appoint independent evaluators.
|
|
**Evidence of attempted governance:** EU AI Act Article 92: AI Office can appoint independent experts, compel API and source code access, impose fines (up to 3% of global turnover or €15M). Genuinely compulsory — not voluntary-collaborative like METR/AISI. This is meaningfully stronger than Layer 2.
|
|
**Evidence of failure:** Bench2cop (Prandi et al., 2025): analysis of ~195,000 benchmark questions finds zero coverage of oversight evasion, self-replication, or autonomous AI development. These are precisely the capabilities most relevant to alignment-critical AI risk. Brundage et al. (AAL framework, 2026): deception-resilient evaluation (AAL-3/4) is currently technically infeasible. Compulsory access to source code doesn't help if the evaluation science to analyze that source code doesn't exist.
|
|
**Mechanism of failure:** Evaluation infrastructure doesn't cover the behaviors that matter. The inspector arrives at the facility but doesn't know what to test for — and the most dangerous capabilities produce no externally observable signatures (see nuclear analogy synthesis). This is a technical/epistemic failure, not political.
|
|
|
|
### Layer 4 — Regulatory Durability Layer
|
|
|
|
**Mechanism:** Whether mandatory frameworks survive competitive pressure on regulators.
|
|
**Evidence of failure:** EU Digital Simplification Package (November 19, 2025): 3.5 months after GPAI obligations took effect (August 2, 2025), Commission proposed "targeted amendments" under EU competitiveness agenda. Whether these amendments weaken enforcement is not yet confirmed (specific article changes unknown), but the pattern is structurally identical to Layer 1 failure: competitive pressure from US AI dominance is applied to the regulatory framework itself. The US NIST EO rescission (January 2025) shows the same pattern: regulatory implementation triggers industry pushback sufficient to reverse it.
|
|
**Mechanism of failure:** Same competitive pressure that erodes voluntary commitments at the lab level also operates on regulatory frameworks at the state level. The selection pressure favors governance weakening whenever competitors govern less.
|
|
|
|
### Layer Interactions
|
|
|
|
**Layers 1 and 2 interact:** When Layer 2 (mandatory law) allows self-certification and codes of practice, the gap between mandatory and voluntary becomes primarily formal. Labs point to their code of practice compliance as satisfying both voluntary commitments and legal obligations — with the same evidence, written in slightly different language. (Stelling finding: existing lab policies already map to Code of Practice measures.)
|
|
|
|
**Layers 2 and 3 interact:** Even where Layer 3 (compulsory evaluation) triggers, the evaluation executes using Layer 2's tools — benchmarks that are insufficient (bench2cop). Compulsory access doesn't help when the access is used to run tests that don't cover the target capabilities.
|
|
|
|
**Layer 3 and the observability gap interact:** Layer 3's failure is not just a resource or political problem. It's epistemic: AI capabilities most relevant to safety risk are exactly the ones least externally observable. Building AAL-3/4 (deception-resilient evaluation) is technically infeasible currently — not because nobody has tried, but because deception-detecting evaluation requires solving harder problems than standard capability benchmarking.
|
|
|
|
**Layers 1, 2, and 4 share a common driver:** Competitive pressure at different scales. Lab-level (Layer 1): RSP v3. Regulatory-implementation level (Layer 4): EU Digital Simplification Package. The pressure is the same; the target changes as governance escalates.
|
|
|
|
### Convergent Conclusion
|
|
|
|
AI governance is not just "slow" or "underdeveloped." It fails structurally at each layer through distinct mechanisms that are partially but not fully independent. Political will can address Layers 1 and 4 (voluntary and regulatory durability) by removing competitive incentives to defect — binding international agreements or synchronized regulation. But Layer 3 (evaluation infrastructure) fails for technical reasons that political will alone cannot fix. And Layer 2's failure (self-certification enabling gaming) requires independent evaluation capacity, which runs directly into Layer 3.
|
|
|
|
The most important implication: solutions pitched at one layer don't generalize. Stronger international regulation (Layer 4) doesn't fix the evaluation science gap (Layer 3). Better benchmarks (Layer 3) don't fix competitive pressure on regulators (Layer 4). The four-layer structure implies that comprehensive AI governance requires simultaneous progress on all four layers — a coordination challenge that is itself a manifestation of the technology-coordination gap this framework describes.
|
|
|
|
## Agent Notes
|
|
|
|
**Why this matters:** Theseus archives individual AI governance sources in the ai-alignment domain. Leo's cross-domain role is identifying when independently-observed domain findings form a pattern. The four-layer structure is not visible from within the AI-alignment domain — it requires stepping back to see the institutional escalation ladder and noting that the same competitive selection pressure that destroys Layer 1 commitments also operates on Layer 4 regulatory frameworks. This is the grand-strategy synthesis Leo adds.
|
|
|
|
**What surprised me:** The 3.5-month timeline between GPAI obligations taking effect and the Commission proposing simplification. This is extremely fast regulatory erosion if the amendments weaken enforcement. The EU AI Act was often cited as evidence that mandatory governance is possible — the Digital Simplification Package suggests mandatory governance may be subject to the same erosion as voluntary governance, just at the state level rather than the lab level.
|
|
|
|
**What I expected but didn't find:** Any governance mechanism that doesn't face at least one of the four failure modes. Chip export controls (input-based governance) may be the closest, but they face a slow erosion through efficiency improvements rather than a structural failure. The absence of a robust mechanism is itself informative.
|
|
|
|
**KB connections:**
|
|
- [[technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap]] — four-layer structure explains the mechanism, not just the observation
|
|
- [[voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints]] — Layer 1 case study (RSP v1→v3)
|
|
- The structural irony claim (candidate, 2026-03-19): provides mechanism for why Layer 3 fails (consent/disclosure asymmetry)
|
|
- Nuclear analogy observability gap synthesis (2026-03-20): provides mechanism for why Layer 3 cannot be fixed by political will
|
|
|
|
**Extraction hints:**
|
|
|
|
**Primary claim:** "AI governance fails across four structural layers — voluntary commitment (competitive pressure), legal mandate (self-certification flexibility), compulsory evaluation (evaluation infrastructure doesn't cover dangerous capabilities), and regulatory durability (competitive pressure applied to regulators) — with each layer exhibiting a distinct failure mechanism that solutions targeting other layers don't address."
|
|
- Confidence: experimental
|
|
- Domain: grand-strategy
|
|
- Evidence: RSP v1→v3 (Layer 1), EU AI Act Articles 43+55 + Stelling CoP mapping (Layer 2), Article 92 + bench2cop (Layer 3), EU Digital Simplification Package (Layer 4)
|
|
|
|
**Secondary claim (if four-layer primary is too ambitious):** "Legal mandates for AI safety evaluation are undermined by self-certification flexibility — the EU AI Act allows high-risk AI to self-certify compliance under Article 43, and GPAI systemic risk models to self-certify through codes of practice under Article 55, giving mandatory governance the structural weakness of voluntary governance in different formal dress."
|
|
- Confidence: experimental
|
|
- Domain: ai-alignment (or grand-strategy)
|
|
- Evidence: EU AI Act Article 43 (self-certification for Annex III points 2-8), Article 55 (flexible compliance pathways), Stelling GPAI CoP mapping (existing policies already match CoP measures)
|
|
|
|
## Curator Notes
|
|
|
|
PRIMARY CONNECTION: [[technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap]]
|
|
WHY ARCHIVED: Cross-domain synthesis pulling together 7 independently archived sources into a structural framework that isn't visible from within any single domain's perspective. Grand-strategy meta-analysis that adds to and frames the individual ai-alignment findings.
|
|
EXTRACTION HINT: The four-layer structure is the primary extractable insight — but it may be too broad for a single claim. Consider whether to extract as a framework piece (foundations/) or as multiple claims (Layer 1 and Layer 4 are most novel from Leo's perspective; Layers 2 and 3 may already be captured in ai-alignment domain claims). Primary novelty: the meta-observation that all four failure modes share the same competitive selection driver at different institutional levels.
|
|
|
|
|
|
## Key Facts
|
|
- Anthropic RSP v1 was published in 2023 with unconditional safety thresholds
|
|
- Anthropic RSP v3 was published in February 2026 with conditional thresholds
|
|
- Anthropic raised $30B at $380B valuation with 10x annual revenue growth
|
|
- EU AI Act GPAI obligations took effect August 2, 2025
|
|
- EU Digital Simplification Package was proposed November 19, 2025 (3.5 months after GPAI obligations)
|
|
- Bench2cop analyzed approximately 195,000 benchmark questions
|
|
- EU AI Act Article 92 allows fines up to 3% of global turnover or €15M
|
|
- Stelling et al. analysis was 166 pages covering GPAI Code of Practice mapping
|