teleo-codex/domains/ai-alignment/technological development draws from an urn containing civilization-destroying capabilities and only preventive governance can avoid black ball technologies.md
m3taversal 381b4f4e48
Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run
theseus: add 5 claims from Bostrom, Russell, Drexler alignment foundations
- What: Phase 3 of alignment research program. 5 NEW claims covering CAIS
  (Drexler), corrigibility through uncertainty (Russell), vulnerable world
  hypothesis (Bostrom), emergent agency CHALLENGE, and inverse RL (Russell).
- Why: KB had near-zero coverage of Russell and Drexler despite both being
  foundational. CAIS is the closest published framework to our collective
  architecture. Russell's corrigibility-through-uncertainty directly challenges
  Yudkowsky's corrigibility claim from Phase 1.
- Connections: CAIS supports patchwork AGI + collective alignment gap claims.
  Emergent agency challenges both CAIS and our collective thesis. Russell's
  off-switch challenges Yudkowsky's corrigibility framing.

Pentagon-Agent: Theseus <46864dd4-da71-4719-a1b4-68f7c55854d3>
2026-04-05 23:55:04 +01:00

5.5 KiB

type domain secondary_domains description confidence source created related
claim ai-alignment
collective-intelligence
Bostrom's Vulnerable World Hypothesis formalizes the argument that some technologies are inherently civilization-threatening and that reactive governance is structurally insufficient — prevention requires surveillance or restriction capabilities that themselves carry totalitarian risk likely Nick Bostrom, 'The Vulnerable World Hypothesis' (Global Policy, 10(4), 2019) 2026-04-05
physical infrastructure constraints on AI scaling create a natural governance window because packaging memory and power bottlenecks operate on 2-10 year timescales while capability research advances in months
voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints
the first mover to superintelligence likely gains decisive strategic advantage because the gap between leader and followers accelerates during takeoff
multipolar failure from competing aligned AI systems may pose greater existential risk than any single misaligned superintelligence

Technological development draws from an urn containing civilization-destroying capabilities and only preventive governance can avoid black ball technologies

Bostrom (2019) introduces the urn model of technological development. Humanity draws balls (inventions, discoveries) from an urn. Most are white (net beneficial) or gray (mixed — benefits and harms). The Vulnerable World Hypothesis (VWH) states that in this urn there is at least one black ball — a technology that, by default, destroys civilization or causes irreversible catastrophic harm.

Bostrom taxonomizes three types of black ball technology:

Type-1 (easy destruction): A technology where widespread access enables mass destruction. The canonical thought experiment: what if nuclear weapons could be built from household materials? The destructive potential already exists in the physics; only engineering difficulty and material scarcity prevent it. If either barrier is removed, civilization cannot survive without fundamentally different governance.

Type-2a (dangerous knowledge): Ideas or information whose mere possession creates existential risk. Bostrom's information hazards taxonomy (2011) provides the formal framework. Some knowledge may be inherently unsafe regardless of the possessor's intentions.

Type-2b (technology requiring governance to prevent misuse): Capabilities that are individually beneficial but collectively catastrophic without coordination mechanisms. This maps directly to multipolar failure from competing aligned AI systems may pose greater existential risk than any single misaligned superintelligence — AI may be a Type-2b technology where individual deployment is rational but collective deployment without coordination is catastrophic.

The governance implications are stark. Bostrom argues that preventing black ball outcomes requires at least one of: (a) restricting technological development (slowing urn draws), (b) ensuring no individual actor can cause catastrophe (eliminating single points of failure), or (c) sufficiently effective global governance including surveillance. He explicitly argues that some form of global surveillance — "turnkey totalitarianism" — may be the lesser evil compared to civilizational destruction. This is his most controversial position.

For AI specifically, the VWH reframes the governance question. physical infrastructure constraints on AI scaling create a natural governance window because packaging memory and power bottlenecks operate on 2-10 year timescales while capability research advances in months — the governance window exists precisely because we haven't yet drawn the AGI ball from the urn. voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints — voluntary coordination fails because black ball dynamics create existential competitive pressure.

The deepest implication: reactive governance is structurally insufficient for black ball technologies. By the time you observe the civilizational threat, prevention is impossible. This is the governance-level equivalent of Yudkowsky's "no fire alarm" thesis — there will be no moment where the danger becomes obvious enough to trigger coordinated action before it's too late. Preventive governance — restricting, monitoring, or coordinating before the threat materializes — is the only viable approach, and it carries its own risks of authoritarian abuse.

Challenges

  • The VWH is unfalsifiable as stated — you cannot prove an urn doesn't contain a black ball. Its value is as a framing device for governance, not as an empirical claim.
  • The surveillance governance solution may be worse than the problem it addresses. History suggests that surveillance infrastructure, once built, is never voluntarily dismantled and is routinely abused.
  • The urn metaphor assumes technologies are "drawn" independently. In practice, technologies co-evolve with governance, norms, and countermeasures. Society adapts to new capabilities in ways the static urn model doesn't capture.
  • Nuclear weapons are arguably a drawn black ball that humanity has survived for 80 years through deterrence and governance — suggesting that even Type-1 technologies may be manageable without totalitarian surveillance.