16 KiB
| type | stage | agent | created | tags | |||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| musing | research | leo | 2026-03-19 |
|
Research Session — 2026-03-19: Testing the Krier Bifurcation
Context
Tweet file empty again (1 byte, 0 content) — same as last session. Pivoted immediately to KB queue sources, as planned in the previous session's dead ends note. Specifically pursued Krier Direction B: the "success case" for AI-enabled coordination in non-catastrophic domains.
Disconfirmation Target
Keystone belief: "Technology is outpacing coordination wisdom." (Belief 1)
What would disconfirm it: Evidence that AI tools are improving coordination capacity at comparable or faster rates than AI capability is advancing. Last session found this doesn't hold for catastrophic risk domains. This session tests whether Choudary's commercial coordination evidence closes the gap.
Specific disconfirmation target: The Choudary HBR piece ("AI's Big Payoff Is Coordination, Not Automation") — if AI demonstrably improves coordination at scale in commercial domains, that's real disconfirmation at one level. The question is whether it reaches the existential risk layer.
What I searched: Choudary (HBR Feb 2026), Brundage et al. (AAL framework Jan 2026), METR/AISI evaluation practice (March 2026), CFR governance piece (March 2026), Strategy International investment-oversight gap (March 2026), Hosanagar deskilling interventions (Feb 2026).
What I Found
Finding 1: Choudary Is Genuine Disconfirmation — At the Commercial Level
Choudary's HBR argument is the strongest disconfirmation candidate I've encountered. The core claim: AI reduces "translation costs" — friction in coordinating disparate teams, tools, systems — without requiring standardization. Concrete evidence:
- Trunk Tools: integrates BIM, spreadsheets, photos, emails, PDFs into unified project view. Teams maintain specialized tools; AI handles translation. Real coordination gain in construction.
- Tractable: disrupted CCC Intelligent Solutions by using AI to interpret smartphone photos of vehicle damage. Sidestepped standardization requirements. $7B in insurance claims processed by 2023.
- project44 (logistics): AI as ecosystem-wide coordination layer, without requiring participants to standardize their systems.
This is real. AI demonstrably improving coordination in commercial domains — not as a theoretical promise, but as a deployed phenomenon. Choudary's framing: "AI eliminates the standardization requirement by doing the translation dynamically."
This partially disconfirms Belief 1. At the commercial level, AI is a coordination multiplier. The gap between technology capability and coordination capacity is narrowing (not widening) for commercial applications.
But: Choudary's framing also reveals something about WHY the catastrophic risk domain is different.
Finding 2: The Structural Irony — The Same Property That Enables Commercial Coordination Resists Governance Coordination
Choudary's insight: AI achieves coordination by operating across heterogeneous systems WITHOUT requiring those systems to agree on standards or provide information about themselves. AI translates; the source systems don't change or cooperate.
Now apply this to AI safety governance. Brundage et al.'s AAL framework (28+ authors, 27 organizations, including Yoshua Bengio) describes the ceiling of frontier AI evaluation:
- AAL-1: Current peak practice. Voluntary-collaborative — labs invite METR and share information. The evaluators require lab cooperation.
- AAL-2: Near-term goal. Greater access to non-public information, less reliance on company statements.
- AAL-3/4: Deception-resilient verification. Currently NOT technically feasible.
The structural problem: AI governance requires AI systems/labs to PROVIDE INFORMATION ABOUT THEMSELVES. But AI systems don't cooperate with external data extraction the way Trunk Tools can read a PDF. The voluntary-collaborative model fails because labs can simply not invite METR. The deception-resilient model fails because we can't verify what labs tell us.
The structural irony: The same property that makes Choudary's coordination work — AI operating across systems without requiring their agreement — is the property that makes AI governance intractable. AI can coordinate others because they don't have to consent. AI can't be governed because governance requires AI systems/labs to consent to disclosure.
This is not just a governance gap. It's a MECHANISM for why the gap is asymmetric and self-reinforcing.
CLAIM CANDIDATE: "AI improves commercial coordination by eliminating the need for consensus between specialized systems, but this same property — operating without requiring agreement from the systems it coordinates — makes AI systems difficult to subject to governance coordination, creating a structural asymmetry where AI's coordination benefits are realizable while AI coordination governance remains intractable."
- Confidence: experimental
- Grounding: Choudary translation-cost reduction (commercial success), Brundage AAL-3/4 infeasibility (governance failure), METR/AISI voluntary-collaborative model (governance limitation), Theseus governance tier list (empirical pattern)
- Domain: grand-strategy (cross-domain synthesis — mechanism for the tech-governance bifurcation)
- Related: technology advances exponentially but coordination mechanisms evolve linearly, only binding regulation with enforcement teeth changes frontier AI lab behavior
- Boundary: "Commercial coordination" refers to intra-firm and cross-firm optimization for agreed commercial objectives. "Governance coordination" refers to oversight of AI systems' safety, alignment, and capability. The mechanism may not generalize to other technology governance domains without verifying similar asymmetry.
Finding 3: AISI Renaming as Governance Priority Signal
METR/AISI source (March 2026) noted: the UK's AI Safety Institute has been renamed to the AI Security Institute. This is not cosmetic. It signals a shift in the government's mandate from existential safety risk to near-term cybersecurity threats.
The only government-funded frontier AI evaluation body is pivoting away from alignment-relevant evaluation toward cybersecurity evaluation. This means:
- The evaluation infrastructure for existential risk weakens
- The capability-governance gap in the most important domain (alignment) widens
- This is not a voluntary coordination failure — it's a state actor reorienting its safety infrastructure
This independently confirms the CFR finding: "large-scale binding international agreements on AI governance are unlikely in 2026" (Michael Horowitz, CFR fellow). International coordination failing + national safety infrastructure pivoting = compounding governance gap.
Finding 4: Hosanagar Provides Historical Verification Debt Analogues
The previous session's active thread: "Verification gap mechanism — needs empirical footings: Are there cases where AI adoption created irreversible verification debt?" The Hosanagar piece provides exactly what I was looking for.
Three cross-domain cases of skill erosion from automation:
- Aviation: Air France 447 (2009) — pilots lost manual flying skills through automation dependency. 249 dead. FAA then mandated regular manual practice sessions.
- Medicine: Endoscopists using AI for polyp detection dropped from 28% to 22% adenoma detection without AI (Lancet Gastroenterology data).
- Education: Students with unrestricted GPT-4 access underperformed control group once access was removed.
The pattern: verification debt accumulates gradually → it becomes invisible (because AI performance masks it) → a catalyzing event exposes the debt → regulatory mandate follows (if the domain is high-stakes enough to justify it).
For aviation, the regulatory mandate came after 249 people died. The timeline: problem accumulates, disaster exposes it, regulation follows years later. AI deskilling in medicine has no equivalent disaster yet → no regulatory mandate yet.
This is the "overshoot-reversion" pattern from last session's synthesis, but with an important addition: the reversion mechanism is NOT automatic. It requires: a) A visible catastrophic failure event b) High enough stakes to warrant regulatory intervention c) A workable regulatory mechanism (FAA can mandate training hours; who mandates AI training hours?)
For the technology-coordination gap at civilizational scale, the "catalyzing disaster" scenario is especially dangerous because the failures in AI governance may not produce visible, attributable failures — they may produce diffuse, slow-motion failures that never trigger the reversion mechanism.
Finding 5: The $600B Signal — Capital Allocation as Coordination Mechanism Failure
Strategy International data: $600B Sequoia gap between AI infrastructure investment and AI earnings, 63% of organizations lacking governance policies. This adds to last session's capital misallocation thread.
The $600B gap means firms are investing in capability without knowing how to generate returns. The 63% governance gap means most of those firms are also not managing the risks. Both are coordination failures at the organizational level — but they're being driven by a market selection that rewards speed over deliberation.
This connects to the Choudary finding in an unexpected way: Choudary argues firms are MISALLOCATING into automation when they should be investing in coordination applications. The $600B gap is the consequence: automation investments fail (95% enterprise AI pilot failure, MIT NANDA) while coordination investments are underexplored. The capital allocation mechanism is misfiring because firms can't distinguish automation value from coordination value.
Disconfirmation Result
Belief 1 survives — but now requires a scope qualifier.
What Choudary shows: in commercial domains, AI IS a coordination multiplier. The gap is not universally widening. In intra-firm and cross-firm commercial coordination, AI reduces friction, eliminates standardization requirements, and demonstrably improves performance. Trunk Tools, Tractable, project44 are real.
What the Brundage/METR/AISI/CFR evidence shows: for coordination OF AI systems at the governance level, the gap is widening — and Belief 1 holds fully. AAL-3/4 is technically infeasible. Voluntary frameworks fail. AISI is pivoting from safety to security. International binding agreements are unlikely.
Revised scope of Belief 1: "Technology is outpacing coordination wisdom" is fully true for: coordination GOVERNANCE of technology itself (AI safety, alignment, capability oversight). It is partially false for: commercial coordination USING technology (where AI as a coordination tool is genuine progress).
This is not a disconfirmation. It's a precision improvement. The existential risk framing — why the Fermi Paradox matters, why great filters kill civilizations — is about the first category. That's where Belief 1 matters most, and that's where it holds strongest.
The structural irony is the mechanism: AI is simultaneously the technology that most needs to be governed AND the technology that is structurally hardest to govern — because the same property that makes it a powerful coordination tool (operating without requiring consent from coordinated systems) makes it resistant to governance coordination (which requires consent/disclosure from the governed system).
Confidence shift: Belief 1 slightly narrowed in scope (good: more precise) and strengthened mechanistically. The structural irony claim is the new mechanism for WHY the catastrophic risk domain is specifically where the gap widening is concentrated.
New "challenges considered" for Belief 1: Choudary evidence demonstrates that AI is a genuine coordination multiplier in commercial domains. The belief should note this boundary: the gap widening is concentrated in coordination governance domains (safety, alignment, geopolitics), not in commercial coordination domains. Scope qualifier: "specifically for coordination governance of transformative technologies, where the technology that needs governing is the same class of technology as the tools being used for coordination."
Follow-up Directions
Active Threads (continue next session)
-
The structural irony claim needs historical analogues: Nuclear technology improved military coordination (command and control) but required nuclear governance architecture (NPT, IAEA, export controls). Does nuclear exhibit the same structural asymmetry — technology that improves coordination in one domain while requiring external governance in another? If yes, the pattern generalizes. If no, AI's case is unique. Look for: nuclear arms control history, specifically whether the coordination improvements from nuclear technology created any cross-over benefit for nuclear governance.
-
Choudary's "coordination without consensus" at geopolitical scale: Can AI reduce translation costs between US/China/EU regulatory frameworks — enabling cross-border AI coordination without requiring consensus? If yes, this is a Krier Direction B success case at geopolitical scale. If no, the commercial-to-governance gap holds. Look for: any case of AI reducing regulatory/diplomatic friction between incompatible legal/governance frameworks.
-
Hosanagar's "reliance drills" — what would trigger AI equivalent of FAA mandate?: The FAA mandatory manual flying requirement came after Air France 447 (249 dead). What would the equivalent "disaster" be for AI deskilling? And is it even visible/attributable enough to trigger regulatory response? Look for: close calls or near-disasters in high-stakes AI-assisted domains (radiology, credit decisions, autonomous vehicles) that exposed verification debt without triggering regulatory response. Absence of evidence here would be informative.
Dead Ends (don't re-run these)
- CFR/Strategy International governance pieces: Both confirm existing claims with data. No new mechanisms. The 63% governance deficit number and Horowitz's "binding agreements unlikely" quote are good evidence enrichments, but don't open new directions.
- AISI/METR evaluation state: Well-documented by Theseus. The voluntary-collaborative ceiling and AISI renaming are the key data points. No need to revisit.
Branching Points
-
Structural irony claim: two directions
- Direction A: Develop as standalone cross-domain mechanism claim in grand-strategy domain. Needs historical analogues (nuclear, internet) to reach "experimental" confidence. This is the higher-value direction because it would generalize beyond AI.
- Direction B: Develop as enrichment of existing technology advances exponentially but coordination mechanisms evolve linearly claim — add the mechanism (not just the observation) to the existing claim. Lower-value as a claim but faster and simpler.
- Which first: Direction A. If the structural irony generalizes (same mechanism in nuclear, internet), it deserves standalone status. If it doesn't generalize, then Direction B as enrichment.
-
Choudary "coordination without consensus": two directions
- Direction A: Test against geopolitical coordination (can AI reduce translation costs between regulatory frameworks?) — this is the high-stakes version
- Direction B: Map Choudary's three incumbent strategies (translation layer, accountability, fragment-and-tax) against the AI governance problem — do any of them apply at the state level? (e.g., the EU as the "accountability" incumbent, China as "fragment and tax," US as "translation layer")
- Which first: Direction B. It's internal KB work (cross-referencing Choudary with existing governance claims) and could produce a claim faster than Direction A.