teleo-codex/inbox/archive/2025-09-26-krier-coasean-bargaining-at-scale.md
m3taversal b64fe64b89
Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run
theseus: 5 claims from ARIA Scaling Trust programme papers
- What: 5 new claims + 6 source archives from papers referenced in
  Alex Obadia's ARIA Research tweet on distributed AGI safety
- Sources: Distributional AGI Safety (Tomašev), Agents of Chaos (Shapira),
  Simple Economics of AGI (Catalini), When AI Writes Software (de Moura),
  LLM Open-Source Games (Sistla), Coasean Bargaining (Krier)
- Claims: multi-agent emergent vulnerabilities (likely), verification
  bandwidth as binding constraint (likely), formal verification economic
  necessity (likely), cooperative program equilibria (experimental),
  Coasean transaction cost collapse (experimental)
- Connections: extends scalable oversight degradation, correlated blind
  spots, formal verification, coordination-as-alignment

Pentagon-Agent: Theseus <B4A5B354-03D6-4291-A6A8-1E04A879D9AC>
2026-03-16 16:46:07 +00:00

29 lines
2.1 KiB
Markdown

---
type: source
title: "Coasean Bargaining at Scale: Decentralization, coordination, and co-existence with AGI"
author: "Seb Krier (Frontier Policy Development, Google DeepMind; personal capacity)"
url: https://blog.cosmos-institute.org/p/coasean-bargaining-at-scale
date_published: 2025-09-26
date_archived: 2026-03-16
domain: ai-alignment
secondary_domains: [collective-intelligence, teleological-economics]
status: processing
processed_by: theseus
tags: [coase-theorem, transaction-costs, agent-governance, decentralization, coordination]
sourced_via: "Alex Obadia (@ObadiaAlex) tweet, ARIA Research Scaling Trust programme"
twitter_id: "712705562191011841"
---
# Coasean Bargaining at Scale
Krier argues AGI agents as personal advocates can dramatically reduce transaction costs, enabling Coasean bargaining at societal scale. Shifts governance from top-down central planning to bottom-up market coordination.
Key arguments:
- Coasean private bargaining has been theoretically sound but practically impossible due to prohibitive transaction costs (discovery, negotiation, enforcement)
- AI agents solve this: instant communication of granular preferences, hyper-granular contracting, automatic verification/enforcement
- Three resulting governance principles: accountability (desires become priced offers), voluntary coalitions (diffuse interests band together at nanosecond speed), continuous self-calibration (rules flex based on live preference streams)
- "Matryoshkan alignment" — nested governance: outer (legal/state), middle (competitive service providers), inner (individual customization)
- Critical limitations acknowledged: wealth inequality, rights allocation remains constitutional/normative, catastrophic risks need state enforcement
- Reframes alignment from engineering guarantees to institutional design
Directly relevant to [[coordination failures arise from individually rational strategies that produce collectively irrational outcomes]] and [[decentralized information aggregation outperforms centralized planning because dispersed knowledge cannot be collected into a single mind]].