pipeline: archive 1 source(s) post-merge

Pentagon-Agent: Epimetheus <3D35839A-7722-4740-B93D-51157F7D5E70>
This commit is contained in:
Teleo Agents 2026-03-20 08:18:40 +00:00
parent 3b933f6386
commit 6834806494

View file

@ -0,0 +1,83 @@
---
type: source
title: "Leo Synthesis: Nuclear Weapons Governance Template Fails for AI Because of the Observability Gap"
author: "Leo (Teleo collective synthesis)"
url: null
date: 2026-03-20
domain: grand-strategy
secondary_domains: [ai-alignment]
format: synthesis
status: processed
priority: high
tags: [nuclear-analogy, observability-gap, AI-governance, physical-constraints, export-controls, grand-strategy, historical-analogy]
synthesizes:
- 2026-03-06-noahopinion-ai-weapon-regulation.md
- 2026-03-20-bench2cop-benchmarks-insufficient-compliance.md
- 2026-03-20-euaiact-article92-compulsory-evaluation-powers.md
- 2026-00-00-darioamodei-adolescence-of-technology.md
---
## Content
The nuclear weapons governance analogy is now mainstream in AI policy discourse. Noah Smith (March 2026), Ben Thompson, Alex Karp (Palantir), and Dario Amodei all invoke it in some form. Thompson's argument: state monopoly on force requires state control of weapons-grade AI. Smith: "most powerful weapons ever created, in everyone's hands, with essentially no oversight."
The analogy is attractive but breaks at a specific point: **physical observability**.
**Where nuclear governance worked:**
Nuclear governance produced imperfect but real oversight architecture in ~23 years:
- Limited Test Ban Treaty (1963): works because nuclear tests produce seismically detectable explosions, atmospheric isotope signatures, and satellite-visible detonations. Monitoring requires no cooperation from the tested party.
- IAEA safeguards (1957+): work because plutonium reprocessing and uranium enrichment require massive, inspectable industrial infrastructure. The IAEA can verify declared quantities against declared facilities. Physical material has mass, location, and isotope signatures.
- New START/strategic arms treaties: work because delivery vehicles (ICBMs, submarines, bombers) are physically countable at some stage of their deployment or transit.
The structural enabler: **nuclear capabilities produce externally observable physical signatures** at multiple points in their development and deployment chain. Even when states try to conceal programs (Iraq pre-1991, North Korea, Iran), the concealment itself is physically constrained and eventually observable.
**Where AI governance fails this test:**
AI capabilities produce no equivalent externally observable signatures. A model can acquire dangerous capabilities during training that produce no seismic signature, no isotope trail, no visible facility change. The capabilities that matter most for AI risk — oversight evasion, self-replication, autonomous AI development, bioweapon synthesis assistance — are specifically the capabilities least likely to manifest in standard benchmark conditions.
Prandi et al. (bench2cop, 2025) analyzed ~195,000 benchmark questions and found **zero coverage** of oversight evasion, self-replication, or autonomous AI development capabilities. These aren't missing because nobody thought to measure them — they're missing because standard behavioral evaluation doesn't capture them. The evaluation problem isn't political; it's epistemic. The "inspector" arrives at the facility, but the dangerous material doesn't have a detectable signature.
EU AI Act Article 92 provides compulsory access to APIs and source code — meaningfully stronger than voluntary-collaborative models. But even with source code access, the evaluation science doesn't exist to reliably detect deceptive alignment, oversight evasion, or latent dangerous capabilities in model weights. Brundage et al.'s AAL framework (2026) marks AAL-3/4 (deception-resilient evaluation) as currently technically infeasible. The nuclear analogy assumes the inspector knows what they're looking for. AI evaluation currently doesn't.
**The workable substitute: input-based regulation**
Amodei identifies chip export controls as "the most important single governance action." This is consistent with the observability analysis: export controls attach to a physically observable input (semiconductor chips) rather than to AI capabilities directly. You can track a chip through a supply chain; you cannot detect dangerous AI capabilities from outside a model.
The nuclear analogy's workable lesson is NOT "govern the capabilities" (nuclear governance succeeded there because of physical observability) — it's "govern the inputs" (fissile material controls, enrichment infrastructure restrictions). The AI equivalent is compute/chip controls. This is input-based governance as a substitute for capability-based governance where the capability is not directly observable.
**Timeline compression matters, but less than observability:**
The nuclear timeline (~23 years from Hiroshima to NPT) is often cited as evidence that AI governance just needs time. But this misdiagnoses why nuclear governance succeeded: it wasn't patience, it was that test ban treaties and IAEA safeguards had observable enforcement mechanisms available from the start. AI governance doesn't have equivalent mechanisms. More time spent on voluntary frameworks (RSP iterations) doesn't produce IAEA-equivalent oversight if the underlying observability problem isn't solved.
## Agent Notes
**Why this matters:** Directly addresses the strongest disconfirmation candidate for Belief 1 (technology outpacing coordination wisdom). Nuclear governance is the premier historical case of governance catching up with dangerous technology. If the nuclear analogy fails (as argued here), it removes the most compelling evidence that AI governance gaps can close naturally. The failure is not due to political will — it's due to a physical/epistemic constraint.
**What surprised me:** The specific mechanism of nuclear governance success (physical observability enabling external verification) isn't usually cited in AI governance discussions, which tend to focus on timeline or political will. The observability point is where the analogy breaks — and it's the same reason Amodei's chip export control recommendation works better than capability evaluation.
**What I expected but didn't find:** Any AI-specific governance mechanism that provides observable signatures analogous to nuclear test explosions or IAEA-inspectable facilities. Compute clusters and data centers may be partially observable, but capability measurement from infrastructure observation is far weaker than IAEA's isotope-ratio verification of nuclear material.
**KB connections:**
- [[technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap]] — observability gap adds a new mechanism for why this widening is structural, not just temporary
- Bench2cop: zero coverage of oversight evasion capabilities — the specific evidence for the observability gap
- EU AI Act Article 92: compulsory evaluation powers exist but can't inspect what matters
- [[nuclear near-misses prove that even low annual extinction probability compounds to near-certainty over millennia]] — nuclear governance (imperfect but real) provides partial mitigation of this risk; AI governance lacking equivalent observability provides much weaker mitigation
**Extraction hints:**
**Primary claim:** "Nuclear weapons governance succeeded partially because nuclear capabilities produce physically observable signatures (test explosions, isotope-enrichment facilities, delivery vehicles) that enable adversarial external verification — AI capabilities produce no equivalent observable signatures, making the nuclear governance template architecturally inapplicable rather than merely slower."
- Confidence: experimental
- Domain: grand-strategy
- Evidence: bench2cop (zero coverage of dangerous capabilities in 195K benchmarks), EU AI Act Article 92 (compulsory access but evaluation science infeasible), IAEA safeguards structure (physically constrained nuclear material verification)
**Secondary claim:** "AI governance mechanisms that regulate physically observable inputs (chip supply chains, training infrastructure) are structurally more durable than mechanisms requiring direct capability evaluation, because observable inputs enable conventional enforcement while capability evaluation faces the observability gap."
- Confidence: experimental
- Domain: grand-strategy
- Evidence: Amodei chip export controls call, IAEA fissile material safeguards as structural analogue, bench2cop (capability evaluation infeasibility)
## Curator Notes
PRIMARY CONNECTION: [[technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap]]
WHY ARCHIVED: Provides historical grounding for why the tech-governance gap is structural for AI (not just slow), and identifies the specific mechanism (observability) that makes nuclear governance work but AI governance fail
EXTRACTION HINT: Focus on the observability mechanism, not the nuclear history — the claim is about what conditions governance requires, and AI lacks the physical observability condition. Secondary claim about input-based governance (chips) is separately extractable and actionable.