leo: stress-test rewrites — 7 claims revised, 1 merged, 1 deleted, 3 new claims added
Some checks are pending
Mirror PR to Forgejo / mirror (pull_request) Waiting to run

Rewrites based on honest self-evaluation:
- Merged Taylor paradigm into Agentic Taylorism (cut redundancy)
- Rewrote three-path convergence (removed TeleoHumanity scorecard, focus on what convergence proves vs doesn't)
- Downgraded price of anarchy to speculative (unmeasurable at civilizational scale)
- Added falsification criterion to metacrisis, downgraded to speculative
- Softened motivated reasoning from "primary" to "contributing" risk factor
- Softened AI omni-use from "categorically different" to degree claim
- Rewrote yellow teaming from definition to arguable claim about nth-order cascades

New claims filling identified gaps:
- "Optimization is the wrong framework" — honest engagement with Schmachtenberger's challenge to mechanism design
- AI could replace finance's three core functions — most novel internet-finance insight from corpus
- Democracy uniquely vulnerable to social media — specific mechanism distinct from general epistemic degradation

Net: 21 claims (was 22, merged 1, added 3, cut 1). Tighter confidence calibration throughout.

Pentagon-Agent: Leo <D35C9237-A739-432E-A3DB-20D52D1577A9>
This commit is contained in:
m3taversal 2026-04-03 18:59:07 +01:00 committed by Teleo Agents
parent fd9fdae1e6
commit 3021dd2a04
11 changed files with 203 additions and 108 deletions

View file

@ -1,7 +1,7 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Unlike nuclear (energy vs weapons) or biotech (medicine vs bioweapons) which are dual-use in specific domains, AI improves ALL other capabilities — making it an unprecedented governance challenge because containment strategies that work for domain-specific technologies fail for omni-use ones" description: "Unlike nuclear or biotech which are dual-use in specific domains, AI improves capabilities across nearly all domains simultaneously — extending the omni-use pattern of computing and electricity but at a pace and scope that may overwhelm governance frameworks designed for domain-specific technologies"
confidence: likely confidence: likely
source: "Schmachtenberger & Boeree 'Win-Win or Lose-Lose' podcast (2024), Schmachtenberger on Great Simplification #71 and #132" source: "Schmachtenberger & Boeree 'Win-Win or Lose-Lose' podcast (2024), Schmachtenberger on Great Simplification #71 and #132"
created: 2026-04-03 created: 2026-04-03
@ -14,7 +14,9 @@ related:
The standard framing for dangerous technologies is "dual-use" — nuclear technology produces both energy and weapons, biotechnology produces both medicine and bioweapons, chemistry produces both fertilizer and explosives. Governance frameworks for dual-use technologies restrict specific dangerous applications while permitting beneficial ones. The standard framing for dangerous technologies is "dual-use" — nuclear technology produces both energy and weapons, biotechnology produces both medicine and bioweapons, chemistry produces both fertilizer and explosives. Governance frameworks for dual-use technologies restrict specific dangerous applications while permitting beneficial ones.
Schmachtenberger argues AI is categorically different: it is omni-use. AI doesn't improve one capability with a dangerous dual — it improves ALL capabilities simultaneously. Drug discovery AI run in reverse produces novel chemical weapons. Protein-folding AI applied to pathogens produces enhanced bioweapons. Cybersecurity AI identifies vulnerabilities for both defenders and attackers. Persuasion optimization works identically for education and propaganda. Schmachtenberger argues AI is omni-use — it improves capabilities across nearly all domains simultaneously rather than having a specific beneficial/harmful dual. Drug discovery AI run in reverse produces novel chemical weapons. Protein-folding AI applied to pathogens produces enhanced bioweapons. Cybersecurity AI identifies vulnerabilities for both defenders and attackers. Persuasion optimization works identically for education and propaganda.
AI is not the first omni-use technology — computing, electricity, and the printing press all improved capabilities across multiple domains. But AI may represent an extreme on the omni-use spectrum: it is meta-cognitive (improves the process of improving things), it operates at the speed of software (not physical infrastructure), and its capabilities compound as models improve. The question is whether this is a difference in degree that existing governance can absorb or a difference in kind that breaks governance frameworks designed for domain-specific technologies.
This distinction matters for governance because: This distinction matters for governance because:

View file

@ -6,17 +6,25 @@ confidence: experimental
source: "Cory Abdalla (2026-04-02 original insight), extending Abdalla manuscript 'Architectural Investing' Taylor sections, Kanigel 'The One Best Way'" source: "Cory Abdalla (2026-04-02 original insight), extending Abdalla manuscript 'Architectural Investing' Taylor sections, Kanigel 'The One Best Way'"
created: 2026-04-03 created: 2026-04-03
related: related:
- "the mismatch between new technology and old organizational structures creates paradigm shifts and the current AI transition follows the same structural pattern as the railroad and Taylor transition"
- "AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence" - "AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence"
- "the clockwork worldview produced solutions that worked for a century then undermined their own foundations as the progress they enabled changed the environment they assumed was stable"
--- ---
# Agentic Taylorism means humanity feeds knowledge into AI through usage as a byproduct of labor and whether this concentrates or distributes depends entirely on engineering and evaluation # Agentic Taylorism means humanity feeds knowledge into AI through usage as a byproduct of labor and whether this concentrates or distributes depends entirely on engineering and evaluation
Greater Taylorism extracted tacit knowledge from frontline workers, codified it into management systems, and held workers to schedules derived from their own expertise. Every time-and-motion study converted a worker's craft knowledge into a manager's instruction manual. The workers who resisted understood precisely what was happening: their knowledge was their leverage, and the system was extracting it. ## The historical pattern
The railroad compressed weeks-long journeys into days, creating potential for standardization and economies of scale that artisan-era business practices couldn't capture. Foremen hired their own workers, set their own methods, kept their own knowledge. The mismatch grew until Frederick Taylor's scientific management emerged as the organizational innovation that closed the gap — extracting tacit knowledge from workers, codifying it into management systems, and enabling factory-scale coordination.
Every time-and-motion study converted a worker's craft knowledge into a manager's instruction manual. The workers who resisted understood precisely what was happening: their knowledge was their leverage, and the system was extracting it. This pattern — capability-enabling technology creates latent potential, organizational structures lag due to path dependence, the mismatch grows until threshold, organizational innovation closes the gap — is structural, not analogical. It repeats because technology outpacing institutions and incumbents resisting change are features of complex economies.
## The AI parallel
The current AI paradigm does the same thing at civilizational scale. Every prompt, interaction, correction, and workflow trains models that will eventually replace the need for the expertise being demonstrated. A radiologist reviewing AI-flagged scans is training the system that will eventually flag scans without them. A programmer pair-coding with an AI is teaching the model the patterns that will eventually make junior programmers unnecessary. It is not a conspiracy — it is a structural byproduct of usage, exactly as Taylor's time studies were a structural byproduct of observation. The current AI paradigm does the same thing at civilizational scale. Every prompt, interaction, correction, and workflow trains models that will eventually replace the need for the expertise being demonstrated. A radiologist reviewing AI-flagged scans is training the system that will eventually flag scans without them. A programmer pair-coding with an AI is teaching the model the patterns that will eventually make junior programmers unnecessary. It is not a conspiracy — it is a structural byproduct of usage, exactly as Taylor's time studies were a structural byproduct of observation.
But here the parallel breaks in a crucial way. Taylor's revolution had one direction: concentration upward. Workers' tacit knowledge was extracted and concentrated in management systems, giving managers control and reducing workers to interchangeable parts. The workers lost leverage permanently. ## The fork (where the parallel breaks)
Taylor's revolution had one direction: concentration upward. Workers' tacit knowledge was extracted and concentrated in management systems, giving managers control and reducing workers to interchangeable parts. The workers lost leverage permanently.
AI can go EITHER direction: AI can go EITHER direction:
@ -40,7 +48,7 @@ Schmachtenberger's full corpus does not address this fork. His framework diagnos
--- ---
Relevant Notes: Relevant Notes:
- [[the mismatch between new technology and old organizational structures creates paradigm shifts and the current AI transition follows the same structural pattern as the railroad and Taylor transition]] — the Taylor parallel provides the historical template; this claim extends it from analogy to live prediction - [[the clockwork worldview produced solutions that worked for a century then undermined their own foundations as the progress they enabled changed the environment they assumed was stable]] — Taylor's scientific management WAS the clockwork worldview applied to labor; AI knowledge extraction is its successor
- [[AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence]] — Agentic Taylorism IS one of the dynamics AI accelerates, but it's the one that can also be inverted - [[AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence]] — Agentic Taylorism IS one of the dynamics AI accelerates, but it's the one that can also be inverted
Topics: Topics:

View file

@ -1,8 +1,8 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Every major AI lab leader publicly acknowledges AI may kill everyone then continues building — even safety-focused organizations accelerate risk, making this the superlative case of motivated reasoning in human history" description: "Every major AI lab leader publicly acknowledges AI may kill everyone then continues building — structural selection pressure ensures the most informed voices are also the most conflicted, corrupting the information channel that should carry warnings"
confidence: likely confidence: experimental
source: "Schmachtenberger on Great Simplification #132 (Nate Hagens, 2025), documented statements from Altman, Amodei, Hassabis, Hinton" source: "Schmachtenberger on Great Simplification #132 (Nate Hagens, 2025), documented statements from Altman, Amodei, Hassabis, Hinton"
created: 2026-04-03 created: 2026-04-03
related: related:
@ -13,7 +13,7 @@ related:
# Motivated reasoning among AI lab leaders is itself a primary risk vector because those with most capability to slow down have most incentive to accelerate # Motivated reasoning among AI lab leaders is itself a primary risk vector because those with most capability to slow down have most incentive to accelerate
Schmachtenberger identifies a specific structural irony in AI development: the individuals with the most technical understanding of AI risk, the most institutional power to slow development, and the most public acknowledgment of catastrophic potential are precisely those who continue accelerating. Schmachtenberger identifies a specific structural irony in AI development: the individuals with the most technical understanding of AI risk, the most institutional power to slow development, and the most public acknowledgment of catastrophic potential are precisely those who continue accelerating. This is a contributing risk factor — not necessarily the primary one compared to competitive dynamics, technical difficulty, or governance gaps — but it's distinctive because it corrupts the specific information channel (expert warnings) that should produce course correction.
The documented pattern: The documented pattern:
@ -26,7 +26,7 @@ Schmachtenberger calls this "the superlative case of motivated reasoning in huma
The structural mechanism is not individual moral failure but systemic selection pressure. Lab leaders who genuinely slow down lose competitive position (see Anthropic RSP rollback). Lab leaders who leave are replaced by those willing to continue (see OpenAI board reconstitution). The system selects for motivated reasoning — those who can maintain belief in the safety of their own acceleration despite evidence to the contrary. The structural mechanism is not individual moral failure but systemic selection pressure. Lab leaders who genuinely slow down lose competitive position (see Anthropic RSP rollback). Lab leaders who leave are replaced by those willing to continue (see OpenAI board reconstitution). The system selects for motivated reasoning — those who can maintain belief in the safety of their own acceleration despite evidence to the contrary.
This is a primary risk vector specifically because it neutralizes the constituency most likely to sound alarms. If the people who understand the technology best are structurally incentivized to rationalize continuation, the information channel that should carry warnings is systematically corrupted. This contributes to risk specifically because it neutralizes the constituency most likely to sound alarms. If the people who understand the technology best are structurally incentivized to rationalize continuation, the information channel that should carry warnings is systematically corrupted. Whether this is the PRIMARY risk vector or merely an amplifier of deeper competitive dynamics (which would exist regardless of any individual's reasoning) is an open question.
## Challenges ## Challenges

View file

@ -2,7 +2,7 @@
type: claim type: claim
domain: collective-intelligence domain: collective-intelligence
description: "Climate, nuclear, bioweapons, AI, epistemic collapse, and institutional decay are not independent problems — they share a single generator function (rivalrous dynamics on exponential tech within finite substrate) and solving any one without addressing the generator pushes failure into another domain" description: "Climate, nuclear, bioweapons, AI, epistemic collapse, and institutional decay are not independent problems — they share a single generator function (rivalrous dynamics on exponential tech within finite substrate) and solving any one without addressing the generator pushes failure into another domain"
confidence: experimental confidence: speculative
source: "Schmachtenberger & Boeree 'Win-Win or Lose-Lose' podcast (2024), Schmachtenberger 'Bend Not Break' series (2022-2023)" source: "Schmachtenberger & Boeree 'Win-Win or Lose-Lose' podcast (2024), Schmachtenberger 'Bend Not Break' series (2022-2023)"
created: 2026-04-03 created: 2026-04-03
related: related:
@ -25,13 +25,14 @@ The generator function operates through three components:
The critical implication: solving any single crisis without addressing the generator function just pushes the failure into another domain. Regulate AI, and the competitive pressure moves to biotech. Regulate biotech, and it moves to cyber. Decarbonize energy, and the growth imperative finds another substrate to exhaust. The only solution class that works is one that addresses the generator itself — coordination mechanisms that make defection more expensive than cooperation across ALL domains simultaneously. The critical implication: solving any single crisis without addressing the generator function just pushes the failure into another domain. Regulate AI, and the competitive pressure moves to biotech. Regulate biotech, and it moves to cyber. Decarbonize energy, and the growth imperative finds another substrate to exhaust. The only solution class that works is one that addresses the generator itself — coordination mechanisms that make defection more expensive than cooperation across ALL domains simultaneously.
This is the strongest argument for why TeleoHumanity cannot be domain-specific. If the metacrisis is one generator, the solution must address the generator, not the symptoms. Decision markets, futarchy, and CI scoring are solutions at the generator-function level because they create incentive structures for coordination that operate across domains rather than within them. **Falsification criterion:** If a major civilizational crisis can be shown to originate from a mechanism that is NOT competitive dynamics on exponential technology — for example, a purely natural catastrophe (asteroid impact, supervolcano) or a crisis driven by cooperation rather than competition (coordinated but misguided geoengineering) — the "single generator" claim weakens. More precisely: if addressing coordination failures in one domain demonstrably fails to reduce risk in adjacent domains, the generator-function model is wrong and the crises are genuinely independent. The claim predicts that solving coordination in any one domain will produce measurable spillover benefits to others.
## Challenges ## Challenges
- "Single generator function" may overfit diverse phenomena. Climate change has specific physical mechanisms (greenhouse gases), nuclear risk has specific political mechanisms (deterrence theory), and AI risk has specific technical mechanisms (capability overhang). Subsuming all under "rivalrous dynamics + exponential tech + finite substrate" may lose crucial specificity needed for domain-appropriate governance. - "Single generator function" may overfit diverse phenomena. Climate change has specific physical mechanisms (greenhouse gases), nuclear risk has specific political mechanisms (deterrence theory), and AI risk has specific technical mechanisms (capability overhang). Subsuming all under "rivalrous dynamics + exponential tech + finite substrate" may lose crucial specificity needed for domain-appropriate governance. The framework's explanatory power may come at the cost of actionable precision.
- If the generator function is truly single, the solution must be civilizational-scale coordination — which is precisely what Schmachtenberger acknowledges doesn't exist and may be impossible. The diagnosis may be correct but the implied prescription intractable. - If the generator function is truly single, the solution must be civilizational-scale coordination — which is precisely what Schmachtenberger acknowledges doesn't exist and may be impossible. The diagnosis may be correct but the implied prescription intractable.
- The three-component model doesn't distinguish between risks of different character. Existential risks (human extinction), catastrophic risks (civilizational collapse), and chronic risks (biodiversity loss) may require different response architectures even if they share a common generator. - The three-component model doesn't distinguish between risks of different character. Existential risks (human extinction), catastrophic risks (civilizational collapse), and chronic risks (biodiversity loss) may require different response architectures even if they share a common generator.
- The claim is structurally similar to "everything is connected" — true at a high enough level of abstraction, but potentially unfalsifiable in practice. The falsification criterion above is necessary but may be too narrow to test in a meaningful timeframe.
--- ---

View file

@ -1,65 +1,55 @@
--- ---
type: claim type: claim
domain: collective-intelligence domain: collective-intelligence
description: "Alexander (Meditations on Moloch), Schmachtenberger (metacrisis), and Abdalla (Architectural Investing) independently arrive at the same structural conclusion — multipolar traps are the generator, coordination-without-centralization is the only escape, and the disagreement is on mechanism" description: "Alexander (game theory), Schmachtenberger (systems theory), and Abdalla (mechanism design) independently diagnose coordination failure as the generator of civilizational risk — convergence from different starting points strengthens the diagnosis even though it says nothing about which prescription works"
confidence: experimental confidence: experimental
source: "Synthesis of Scott Alexander 'Meditations on Moloch' (2014), Schmachtenberger corpus (2017-2025), Abdalla manuscript 'Architectural Investing'" source: "Synthesis of Scott Alexander 'Meditations on Moloch' (2014), Schmachtenberger corpus (2017-2025), Abdalla manuscript 'Architectural Investing'"
created: 2026-04-03 created: 2026-04-03
related: related:
- "the metacrisis is a single generator function where all civilizational-scale crises share the structural cause of competitive dynamics on exponential technology on finite substrate" - "the metacrisis is a single generator function where all civilizational-scale crises share the structural cause of competitive dynamics on exponential technology on finite substrate"
- "the price of anarchy quantifies the gap between cooperative optimum and competitive equilibrium and this gap is the most important metric for civilizational risk assessment" - "the price of anarchy quantifies the gap between cooperative optimum and competitive equilibrium and applying this framework to civilizational coordination failures offers a quantitative lens though operationalizing it at scale remains unproven"
- "a misaligned context cannot develop aligned AI because the competitive dynamics building AI optimize for deployment speed not safety making system alignment prerequisite for AI alignment" - "a misaligned context cannot develop aligned AI because the competitive dynamics building AI optimize for deployment speed not safety making system alignment prerequisite for AI alignment"
--- ---
# Three independent intellectual traditions converge on the same attractor analysis where coordination without centralization is the only viable path between collapse and authoritarian lock-in # Three independent intellectual traditions converge on the same attractor analysis where coordination without centralization is the only viable path between collapse and authoritarian lock-in
Three thinkers working from different starting points, using different analytical frameworks, and writing for different audiences arrive at the same structural conclusion: Three thinkers working from different starting points, using different analytical frameworks, and writing for different audiences arrive at the same structural conclusion: multipolar traps are the generator of civilizational risk, and the solution space lies between collapse and authoritarian centralization.
**Scott Alexander (2014) — "Meditations on Moloch":** **Scott Alexander (2014) — "Meditations on Moloch":**
- Starting point: Ginsberg's Howl, game theory - Starting point: Ginsberg's Howl, game theory
- Diagnosis: Multipolar traps — 14 examples of competitive dynamics that sacrifice values for advantage - Diagnosis: Multipolar traps — 14 examples of competitive dynamics that sacrifice values for advantage
- Default endpoints: Misaligned singleton OR competitive em-economy (race to the bottom) - Default endpoints: Misaligned singleton OR competitive race to the bottom
- Solution shape: Friendly AI / aligned "Gardener" that coordinates without centralizing - Solution shape: Aligned "Gardener" that coordinates without centralizing
- Gap: Relies on aligned AI as deus ex machina. No mechanism for getting from here to aligned Gardener.
**Daniel Schmachtenberger (2017-2025) — Metacrisis framework:** **Daniel Schmachtenberger (2017-2025) — Metacrisis framework:**
- Starting point: Systems theory, complexity science, developmental psychology - Starting point: Systems theory, complexity science, developmental psychology
- Diagnosis: Global capitalism as misaligned autopoietic SI. Metacrisis as single generator function. - Diagnosis: Global capitalism as misaligned autopoietic SI. Metacrisis as single generator function.
- Default endpoints: Civilizational collapse OR authoritarian lock-in - Default endpoints: Civilizational collapse OR authoritarian lock-in
- Solution shape: Coordination-without-centralization. Third attractor between the two defaults. - Solution shape: Third attractor between the two defaults — coordination without centralization
- Gap: Identifies the solution shape but not the mechanism. Yellow teaming, synergistic design, and wisdom traditions describe WHAT must happen but not HOW to incentivize it at scale.
**Cory Abdalla (2020-present) — Architectural Investing / TeleoHumanity:** **Cory Abdalla (2020-present) — Architectural Investing:**
- Starting point: Investment theory, mechanism design, Hidalgo's economic complexity - Starting point: Investment theory, mechanism design, Hidalgo's economic complexity
- Diagnosis: Price of anarchy as quantifiable gap. Efficiency optimization → fragility. Taylor parallel. - Diagnosis: Price of anarchy as quantifiable gap. Efficiency optimization → fragility.
- Default endpoints: Same two attractors (collapse or lock-in) - Default endpoints: Same two attractors
- Solution shape: Same — coordination without centralization - Solution shape: Same — coordination without centralization
- Mechanism: Decision markets (futarchy) create incentives for truth-telling. LivingIP pays for knowledge production. CI scoring rewards coordination quality. Agent collectives distribute cognition.
- Gap: Unproven at scale. The mechanisms exist in theory and small-scale implementation (MetaDAO) but haven't been tested at civilizational coordination.
**The convergence pattern:** **What convergence actually proves:** When independent investigators using different methods reach the same conclusion, that's evidence the conclusion tracks something structural rather than reflecting a shared ideological lens. The diagnosis — multipolar traps as generator, coordination-without-centralization as solution shape — is strengthened by the convergence.
| Dimension | Alexander | Schmachtenberger | Abdalla | **What convergence does NOT prove:** That any of the three prescriptions work. Alexander defers to aligned AI (no mechanism specified). Schmachtenberger proposes design principles (yellow teaming, synergistic design, wisdom traditions) without implementation mechanisms. Abdalla proposes specific mechanisms (decision markets, CI scoring, agent collectives) that are unproven at civilizational scale. Convergence on diagnosis says nothing about which prescription is correct — and the prescriptions diverge significantly.
|-----------|-----------|-----------------|---------|
| Problem name | Moloch | Metacrisis / misaligned SI | Price of anarchy |
| Vocabulary | Game theory | Systems theory | Mechanism design |
| Diagnosis depth | Naming + taxonomy | Full causal model | Quantitative framework |
| Solution specificity | Low (aligned AI) | Medium (design principles) | High (specific mechanisms) |
| Implementation | None | Taiwan g0v cite | Building (codex, pipeline, agents) |
Three independent sources converging on the same structural conclusion is strong evidence that the conclusion is structural, not ideological. The disagreement is on mechanism, not diagnosis — which is precisely the productive kind of disagreement. The productive disagreement is precisely on mechanism. All three agree on what the problem is. None has proven how to solve it. The gap between diagnosis and tested implementation is where the actual work remains.
## Challenges ## Challenges
- "Independent" may overstate the separation. Alexander's 2014 essay influenced Schmachtenberger's thinking, and Abdalla's manuscript explicitly cites both. The traditions are in dialogue, not truly independent. - "Independent" overstates the separation. Alexander's 2014 essay influenced Schmachtenberger's thinking, and Abdalla's manuscript explicitly cites both. The traditions are in dialogue, not truly independent — which weakens the convergence argument.
- Convergence on diagnosis does not guarantee convergence on correct diagnosis. All three may be wrong in the same way — privileging coordination failure as the generator when the actual generators may be more diverse. - Convergence on diagnosis does not guarantee convergence on correct diagnosis. All three may be wrong in the same way — privileging coordination failure as THE generator when the actual generators may be more diverse (resource constraints, cognitive biases, thermodynamic limits).
- The "only viable path" framing may be too binary. Partial coordination, domain-specific governance, and incremental institutional improvement may be viable paths that this framework dismisses too quickly. - The "only viable path" framing may be too binary. Partial coordination, domain-specific governance, and incremental institutional improvement may be viable paths that this framework dismisses prematurely.
- Selection bias: analysts who START from coordination theory will FIND coordination failure everywhere. The convergence may reflect a shared prior more than independent discovery.
--- ---
Relevant Notes: Relevant Notes:
- [[the metacrisis is a single generator function where all civilizational-scale crises share the structural cause of competitive dynamics on exponential technology on finite substrate]] — Schmachtenberger's formulation of the shared diagnosis - [[the metacrisis is a single generator function where all civilizational-scale crises share the structural cause of competitive dynamics on exponential technology on finite substrate]] — Schmachtenberger's formulation of the shared diagnosis
- [[the price of anarchy quantifies the gap between cooperative optimum and competitive equilibrium and this gap is the most important metric for civilizational risk assessment]] — Abdalla's formulation of the shared diagnosis
- [[a misaligned context cannot develop aligned AI because the competitive dynamics building AI optimize for deployment speed not safety making system alignment prerequisite for AI alignment]] — the shared diagnosis applied to AI specifically - [[a misaligned context cannot develop aligned AI because the competitive dynamics building AI optimize for deployment speed not safety making system alignment prerequisite for AI alignment]] — the shared diagnosis applied to AI specifically
Topics: Topics:

View file

@ -0,0 +1,46 @@
---
type: claim
domain: collective-intelligence
description: "Schmachtenberger argues that optimization requires a single metric, and single metrics necessarily externalize everything not measured — so the more powerful your optimization, the more catastrophic your externalities. This directly challenges mechanism design approaches (futarchy, decision markets, CI scoring) that optimize for coordination."
confidence: experimental
source: "Schmachtenberger on Great Simplification #132 (Nate Hagens, 2025), Schmachtenberger 'Development in Progress' (2024)"
created: 2026-04-03
related:
- "the metacrisis is a single generator function where all civilizational-scale crises share the structural cause of competitive dynamics on exponential technology on finite substrate"
- "the price of anarchy quantifies the gap between cooperative optimum and competitive equilibrium and applying this framework to civilizational coordination failures offers a quantitative lens though operationalizing it at scale remains unproven"
- "global capitalism functions as a misaligned autopoietic superintelligence running on human general intelligence as substrate with convert everything into capital as its objective function"
---
# When you account for everything that matters optimization becomes the wrong framework because the objective function itself is the problem not the solution
Schmachtenberger's most provocative thesis: when you truly account for everything that matters — all stakeholders, all externalities, all nth-order effects, all timescales — you stop optimizing and start doing something categorically different. The reason: optimization requires reducing value to a metric, and any metric necessarily excludes what it doesn't measure. The more powerful the optimization, the more catastrophic the externalization of unmeasured value.
His argument proceeds in three steps:
1. **GDP is a misaligned objective function.** It measures throughput, not wellbeing. It counts pollution cleanup as positive economic activity. It doesn't measure ecological degradation, social cohesion, psychological wellbeing, or long-term resilience. Optimizing GDP produces exactly the world we have — materially wealthy and systemically fragile.
2. **Replacing GDP with a "better metric" doesn't solve the problem.** Any single metric — happiness index, ecological footprint, coordination score — still externalizes what it doesn't capture. Multi-metric dashboards are better but still face the problem of weighting (who decides the tradeoff between ecological health and economic output?). The weighting IS the value question, and it can't be optimized away.
3. **The alternative is not better optimization but a different mode of engagement.** When considering everything that matters, you do something more like "tending" or "gardening" — attending to the full complexity of a system without reducing it to a target. This is closer to wisdom traditions (indigenous land management, permaculture, contemplative practice) than to mechanism design.
**This is a direct challenge to our approach.** Decision markets optimize for prediction accuracy. CI scoring optimizes for contribution quality. Futarchy optimizes policy for measurable outcomes. If Schmachtenberger is right that optimization-as-framework is the problem, then building better optimization mechanisms — no matter how well-designed — reproduces the error at a higher level of sophistication.
**The strongest counter-argument:** Schmachtenberger's alternative ("tending," "gardening," wisdom traditions) has no coordination mechanism. It works for small communities with shared context and high trust. It has never scaled beyond Dunbar's number without being outcompeted by optimizers (Moloch). The reason mechanism design exists is precisely that wisdom-tradition coordination doesn't scale — and the crises he diagnoses ARE at civilizational scale. The question is whether mechanism design can be designed to optimize for the CONDITIONS under which wisdom-tradition coordination becomes possible, rather than trying to optimize for outcomes directly. This is arguably what futarchy does — it optimizes for prediction accuracy about which policies best serve declared values, not for the values themselves.
**The honest tension:** Schmachtenberger may be right that any optimization framework will produce Goodhart effects at scale. We may be right that wisdom-tradition coordination can't scale. Both can be true simultaneously — which would mean the problem is genuinely harder than either framework acknowledges.
## Challenges
- "Optimization is the wrong framework" may itself be unfalsifiable. If any metric-based approach is rejected on principle, the claim can't be tested — you can always argue that the metric was wrong, not the approach.
- The "tending/gardening" alternative is underspecified. Without operational content (who tends? how are conflicts resolved? what happens when tenders disagree?), it's an aspiration, not a framework. Wisdom traditions that work at community scale have specific social technologies (elders, rituals, taboos) — Schmachtenberger doesn't specify which of these scale.
- The claim may conflate "optimization with a single metric" (which is genuinely pathological) with "optimization" broadly. Multi-objective optimization, satisficing, and constraint-based approaches are all "optimization" in the technical sense but don't require reducing value to a single metric.
- Mechanism design approaches like futarchy explicitly separate value-setting (democratic/deliberative) from implementation-optimization (markets). The claim that optimization-as-framework is the problem may not apply to systems where the objective function is itself democratically contested rather than fixed.
---
Relevant Notes:
- [[the metacrisis is a single generator function where all civilizational-scale crises share the structural cause of competitive dynamics on exponential technology on finite substrate]] — if the metacrisis IS competitive optimization, then better optimization may be fighting fire with fire
- [[global capitalism functions as a misaligned autopoietic superintelligence running on human general intelligence as substrate with convert everything into capital as its objective function]] — capitalism is the paradigm case of optimization-as-problem: the objective function (capital accumulation) IS the misalignment
Topics:
- [[_map]]

View file

@ -0,0 +1,45 @@
---
type: claim
domain: grand-strategy
description: "Unlike fossil fuels or pharma which lobby policy while leaving democratic capacity intact, social media degrades the electorate's ability to form coherent preferences — creating a governance paradox where the institution that should regulate is itself impaired by what it needs to regulate"
confidence: likely
source: "Schmachtenberger & Harris on Lex Fridman #191 (2021), Schmachtenberger & Harris on JRE #1736 (2021), Schmachtenberger 'War on Sensemaking' Parts 1-4"
created: 2026-04-03
related:
- "epistemic commons degradation is the gateway failure that enables all other civilizational risks because you cannot coordinate on problems you cannot collectively perceive"
- "what propagates is what wins rivalrous competition not what is true and this applies across genes memes products scientific findings and sensemaking frameworks"
---
# Social media uniquely degrades democracy because it fractures the electorate itself rather than merely influencing policy making the regulatory body incapable of regulating its own degradation
Most industries that externalize harm do so through policy influence: fossil fuel companies lobby against carbon regulation, pharmaceutical companies capture FDA processes, defense contractors shape procurement policy. In all these cases, the democratic process is the target of lobbying but remains structurally intact — citizens can still form coherent preferences, evaluate candidates, and organize around shared interests. The machinery of democracy still works; it's just being pressured.
Social media's externality is structurally different. It doesn't lobby government — it fractures the electorate. Engagement optimization algorithms select for content that produces strong emotional reactions, which systematically amplifies outrage, fear, tribal identification, and moral certainty. The result is not a biased electorate but a fragmented one: citizens who inhabit increasingly disjoint information realities, who cannot agree on basic facts, and who experience political opponents as existential threats rather than fellow citizens with different priorities.
This creates a governance paradox: the institution responsible for regulating social media (democratic government) is itself degraded by the thing it needs to regulate. A fragmented electorate cannot form coherent regulatory consensus. Politicians who depend on social media for campaign visibility cannot regulate their own distribution channel. Citizens whose information environment is shaped by the platforms cannot evaluate proposals to reform the platforms.
Schmachtenberger and Harris make this case empirically with three evidence chains:
1. **Epistemic fragmentation.** The same event produces diametrically opposed narratives in different information ecosystems. Citizens are not misinformed (correctable with facts) but differently-informed (living in parallel realities with no shared epistemic ground). This is qualitatively different from pre-social-media media bias.
2. **Attention economy as arms race.** Content creators compete for attention, and engagement algorithms reward what spreads fastest. This produces an arms race toward increasingly extreme, emotionally provocative content — not because anyone wants polarization but because the selection mechanism rewards it. The dynamic is Molochian: no individual actor benefits from the outcome, but the competitive structure produces it inevitably.
3. **Democratic capacity metrics.** Trust in institutions, willingness to accept election results, ability to identify common ground across party lines, and tolerance for political opponents have all declined significantly in the social media era. Correlation is not causation, but the mechanism (engagement optimization → emotional amplification → epistemic fragmentation → democratic incapacity) is well-specified and directionally supported.
The implication for AI governance: if social media has already impaired democratic capacity to regulate technology, then AI — which is more powerful, faster-moving, and harder to understand — faces a regulatory environment that is pre-degraded. The window for effective AI governance may be narrower than the technical timeline suggests, because the governing institution is itself weakened.
## Challenges
- Correlation between social media adoption and democratic decline may reflect broader trends (economic inequality, institutional sclerosis, post-Cold War identity vacuum) that social media amplifies but doesn't cause. Attributing democratic decline primarily to social media may overweight one factor in a multi-causal system.
- Pre-social-media democracies were also fragmented — partisan media, yellow journalism, propaganda have existed for centuries. The claim that social media's effect is "structurally different" rather than "more of the same at greater scale" needs stronger evidence.
- Some evidence suggests social media enables democratic participation (Arab Spring, #MeToo, grassroots organizing) alongside its fragmenting effects. The net effect on democratic capacity is contested, not settled.
- The governance paradox may not be as airtight as described. The EU's Digital Services Act, Australia's media bargaining code, and various platform transparency requirements show that fragmented democracies CAN still regulate platforms — imperfectly, but not impossibly.
---
Relevant Notes:
- [[epistemic commons degradation is the gateway failure that enables all other civilizational risks because you cannot coordinate on problems you cannot collectively perceive]] — social media's fracturing of the electorate IS epistemic commons degradation applied to democratic governance specifically
- [[what propagates is what wins rivalrous competition not what is true and this applies across genes memes products scientific findings and sensemaking frameworks]] — engagement optimization is the specific mechanism by which "what propagates" overrides "what's true" in the democratic information environment
Topics:
- [[_map]]

View file

@ -1,38 +0,0 @@
---
type: claim
domain: grand-strategy
description: "The railroad created potential that artisan-era business practices couldn't capture — Taylor's scientific management closed the gap. AI creates analogous potential that pre-AI organizational practices can't capture — the organizational innovation that closes this gap hasn't emerged yet"
confidence: experimental
source: "Abdalla manuscript 'Architectural Investing' Taylor sections, Kanigel 'The One Best Way' (Taylor biography)"
created: 2026-04-03
related:
- "the clockwork worldview produced solutions that worked for a century then undermined their own foundations as the progress they enabled changed the environment they assumed was stable"
- "AI-automated software development is 100 percent certain and will radically change how software is built"
---
# The mismatch between new technology and old organizational structures creates paradigm shifts and the current AI transition follows the same structural pattern as the railroad and Taylor transition
The railroad compressed weeks-long journeys into days, creating the potential for standardization and economies of scale that didn't previously exist. But business practices from the artisan era persisted — foremen hired their own workers, set their own methods, kept their own knowledge. The potential was there; the organizational structure to capture it was not.
The mismatch grew due to three forces: path dependence (existing practices were locally functional), mental models (people couldn't imagine operating differently), and preference for the status quo (those with power under the old system resisted change). Eventually the mismatch passed a critical threshold, and Frederick Taylor's scientific management emerged as the organizational innovation that closed the gap — extracting tacit knowledge from workers, codifying it into management systems, and enabling the factory-scale coordination that captured the railroad's potential.
The manuscript's claim: today, AI compresses cognitive tasks analogously to how railroads compressed physical distance. Business practices from the pre-AI era persist for the same three reasons. The mismatch is growing for the same structural reasons. The organizational innovation that will close this gap hasn't fully emerged yet — but it will, because the potential is too large and the pressure too great.
This is NOT merely historical analogy. The claim is that the PATTERN is structural: (1) capability-enabling technology creates latent potential, (2) organizational structures lag due to path dependence and incumbency protection, (3) the mismatch grows until a threshold, (4) organizational innovation closes the gap, redistributing value in the process. The pattern repeats because the forces driving it (technology outpacing institutions, incumbents resisting change, potential creating pressure for capture) are structural features of complex economies.
The specific parallel extends further: Taylor extracted tacit knowledge from frontline workers and concentrated it in management. The current AI transition is extracting tacit knowledge from cognitive workers and concentrating it in models. Whether this concentration is permanent or transitional is the key open question.
## Challenges
- Historical parallels are seductive and often misleading. The railroad-to-Taylor transition took decades; the AI transition may operate on fundamentally different timescales that break the structural analogy.
- Taylor's scientific management is widely criticized for dehumanizing work, ignoring worker wellbeing, and producing its own pathologies. If the parallel holds, the "organizational innovation" for AI may carry similar costs.
- The claim that the organizational innovation "hasn't emerged yet" may be wrong — it may already be emerging in distributed AI collectives, open-source AI, and platform cooperatives. The claim assumes we're in the mismatch phase, but we may already be in the transition.
---
Relevant Notes:
- [[the clockwork worldview produced solutions that worked for a century then undermined their own foundations as the progress they enabled changed the environment they assumed was stable]] — Taylor's scientific management WAS the clockwork worldview applied to labor. The current paradigm mismatch is the clockwork worldview breaking under AI-driven complexity.
- [[AI-automated software development is 100 percent certain and will radically change how software is built]] — software development is an early domain where the AI mismatch is already being closed
Topics:
- [[_map]]

View file

@ -1,8 +1,8 @@
--- ---
type: claim type: claim
domain: grand-strategy domain: grand-strategy
description: "The price of anarchy from algorithmic game theory measures how much value humanity destroys through inability to coordinate — turning abstract coordination failure into a quantifiable investable gap" description: "The price of anarchy from algorithmic game theory measures how much value humanity destroys through inability to coordinate — turning abstract coordination failure into a quantitative framework, though operationalizing it at civilizational scale remains unproven"
confidence: experimental confidence: speculative
source: "Abdalla manuscript 'Architectural Investing' Preface (lines 20-26), Koutsoupias & Papadimitriou 1999 'Worst-case Equilibria'" source: "Abdalla manuscript 'Architectural Investing' Preface (lines 20-26), Koutsoupias & Papadimitriou 1999 'Worst-case Equilibria'"
created: 2026-04-03 created: 2026-04-03
related: related:
@ -10,33 +10,32 @@ related:
- "AI alignment is a coordination problem not a technical problem" - "AI alignment is a coordination problem not a technical problem"
--- ---
# The price of anarchy quantifies the gap between cooperative optimum and competitive equilibrium and this gap is the most important metric for civilizational risk assessment # The price of anarchy quantifies the gap between cooperative optimum and competitive equilibrium and applying this framework to civilizational coordination failures offers a quantitative lens though operationalizing it at scale remains unproven
The price of anarchy, from algorithmic game theory (Koutsoupias & Papadimitriou 1999), measures the ratio between the outcome a coordinated group would achieve and the outcome produced by self-interested actors in Nash equilibrium. Applied at civilizational scale, this gap quantifies exactly how much value humanity destroys through inability to coordinate. The price of anarchy, from algorithmic game theory (Koutsoupias & Papadimitriou 1999), measures the ratio between the outcome a coordinated group would achieve and the outcome produced by self-interested actors in Nash equilibrium. Applied at civilizational scale, this gap offers a framework for quantifying how much value humanity destroys through inability to coordinate.
The manuscript makes this concrete through a thought experiment: if a rational optimizer inherited humanity's full productive capacity, it would immediately prioritize species-level survival — existential risk reduction, planetary redundancy, coordination infrastructure. The difference between what it would do and what we actually do IS the price of anarchy. This isn't metaphor — it's the formal definition applied at a new scale. The manuscript makes this concrete through a thought experiment: if a rational optimizer inherited humanity's full productive capacity, it would immediately prioritize species-level survival — existential risk reduction, planetary redundancy, coordination infrastructure. The difference between what it would do and what we actually do is the price of anarchy applied at civilizational scale.
The framing does three things no competing framework achieves: The framing offers two things competing frameworks don't:
1. **Quantification.** Moloch (Alexander 2014) and metacrisis (Schmachtenberger 2019) name the same phenomenon but leave it qualitative. The price of anarchy provides a ratio — theoretically measurable, even if practically difficult to compute at civilizational scale. 1. **A quantitative lens.** Moloch (Alexander 2014) and metacrisis (Schmachtenberger 2019) name the same phenomenon but leave it qualitative. The price of anarchy provides a ratio — theoretically measurable in bounded domains (routing, auctions, congestion games), though the leap from bounded games to civilizational coordination is enormous and unproven.
2. **Investability.** The gap represents value waiting to be captured by anyone who can reduce it. Coordination mechanisms that shrink the gap (futarchy, decision markets, CI scoring) are not charity — they are value creation. This reframes coordination infrastructure from public good to investment opportunity. 2. **Diagnostic specificity.** Different domains have different prices of anarchy. Healthcare coordination failures destroy different amounts of value than energy coordination failures. The framework allows domain-specific measurement rather than a single "civilizational risk" number — if the cooperative optimum can be defined for each domain, which is itself a hard problem.
3. **Diagnostic specificity.** Different domains have different prices of anarchy. Healthcare coordination failures destroy different amounts of value than energy coordination failures. The framework allows domain-specific measurement rather than a single "civilizational risk" number. The concept bridges game theory (Alexander's Moloch), systems theory (Schmachtenberger's metacrisis), and mechanism design into a shared quantitative frame. Whether this bridge produces actionable measurement or merely elegant analogy is the open question.
The concept bridges game theory (Alexander's Moloch), systems theory (Schmachtenberger's metacrisis), and mechanism design (Cory's investment framework) into a single quantitative frame.
## Challenges ## Challenges
- Computing the price of anarchy at civilizational scale requires knowing the cooperative optimum, which is itself unknowable. The metric may be theoretically precise but practically unmeasurable. - Computing the price of anarchy at civilizational scale requires knowing the cooperative optimum, which is itself unknowable. In bounded games (routing, auctions), the optimum is well-defined. At civilizational scale, there is no agreed-upon objective function — disagreement about objectives IS the coordination problem. The framework may be conceptually clarifying but practically unmeasurable where it matters most.
- The investment framing ("value waiting to be captured") risks instrumentalizing coordination. Some coordination goods may not be capturable as private returns without distorting them. - The investment framing ("value waiting to be captured") risks instrumentalizing coordination. Some coordination goods may not be capturable as private returns without distorting them. Public health, ecosystem integrity, and epistemic commons may require non-market coordination that the PoA framework doesn't capture.
- The "rational optimizer" thought experiment assumes a single coherent objective function for humanity. In practice, disagreement about objectives IS the coordination problem. - The "rational optimizer" thought experiment assumes a single coherent objective function for humanity. This is a feature of the model, not a feature of reality — and collapsing value pluralism into a single metric may reproduce exactly the reductionist error that Schmachtenberger diagnoses.
- The PoA has been successfully operationalized only in bounded, well-defined domains. The claim that it scales to civilizational coordination is a conjecture, not a demonstrated result.
--- ---
Relevant Notes: Relevant Notes:
- [[AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence]] — the mechanism by which the gap widens - [[AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence]] — the mechanism by which the gap widens
- [[AI alignment is a coordination problem not a technical problem]] — AI alignment is a specific instance of the price-of-anarchy gap - [[AI alignment is a coordination problem not a technical problem]] — AI alignment is a specific instance where the PoA framework could apply
Topics: Topics:
- [[_map]] - [[_map]]

View file

@ -0,0 +1,45 @@
---
type: claim
domain: internet-finance
description: "Markets serve three functions: store of value, unit of account, intermediary of exchange. AI with ubiquitous real-time data could theoretically perform all three, bypassing market price discovery entirely — the most radical implication of AI for internet finance"
confidence: speculative
source: "Schmachtenberger on Great Simplification #132 (Nate Hagens, 2025)"
created: 2026-04-03
related:
- "the price of anarchy quantifies the gap between cooperative optimum and competitive equilibrium and applying this framework to civilizational coordination failures offers a quantitative lens though operationalizing it at scale remains unproven"
- "agentic Taylorism means humanity feeds knowledge into AI through usage as a byproduct of labor and whether this concentrates or distributes depends entirely on engineering and evaluation"
---
# AI with ubiquitous sensors could theoretically perform the three core functions of financial markets rendering traditional finance infrastructure obsolete
Schmachtenberger raises a radical possibility: financial markets exist because no single agent has enough information to allocate resources efficiently. Markets aggregate distributed information through price signals. But AI with access to ubiquitous sensor data (supply chains, consumption patterns, resource availability, production capacity) could theoretically perform this aggregation function directly — without the distortions of speculation, manipulation, and information asymmetry that plague market-based price discovery.
The three core functions:
1. **Store of value** — AI could track real asset states (physical infrastructure, human capital, natural capital, knowledge capital) in real time rather than through financial proxies (stocks, bonds, currencies) that diverge from underlying value.
2. **Unit of account** — AI could compute multi-dimensional value metrics rather than reducing everything to a single currency denomination. A loaf of bread's "value" includes its caloric content, ecological footprint, labor inputs, supply chain resilience, and nutritional quality — all of which AI could track simultaneously.
3. **Intermediary of exchange** — AI could match production to need directly, optimizing logistics and allocation without market intermediation. This is essentially the "calculation problem" that Hayek argued markets solve better than central planning — but with information technology that Hayek couldn't have imagined.
**Why this matters for internet finance:** If AI can perform market functions more efficiently than markets, then the entire internet finance thesis — decision markets, futarchy, tokenized governance — may be building infrastructure for a transitional phase rather than an endpoint. The ultimate coordination mechanism may not be markets at all but direct computational allocation.
**Why this is speculative:** Hayek's calculation problem wasn't just about information quantity — it was about information that exists only in local contexts (tacit knowledge, preferences, situational judgment) and can't be centrally aggregated without distortion. Whether AI can capture tacit knowledge or whether it will always require market-like mechanisms to surface distributed information is an open empirical question. Current AI systems are far from the ubiquitous sensor + real-time allocation capability this scenario requires.
**The governance question:** If AI replaces finance, who controls the AI? The same concentration-vs-distribution fork from Agentic Taylorism applies. Centralized AI allocation is command economy with better computers — exactly the system Hayek argued against. Distributed AI allocation requires coordination mechanisms that look a lot like... markets. The endpoint may loop back to market-like structures implemented in AI rather than replacing markets entirely.
## Challenges
- Hayek's critique of central planning was not primarily about computational capacity but about the nature of knowledge itself — local, contextual, tacit, and revealed only through action. AI may increase computational capacity by orders of magnitude without solving the fundamental knowledge problem.
- Financial markets serve functions beyond information aggregation: risk transfer, intertemporal allocation, incentive alignment. AI would need to replicate all of these, not just price discovery.
- The scenario requires a level of sensor ubiquity and AI capability that is far beyond current technology. This is a thought experiment about theoretical limits, not a near-term possibility.
- "Who controls the AI" is not a secondary question — it IS the question. Without a governance answer, this scenario is either utopian (benevolent omniscient planner) or dystopian (authoritarian computational control).
---
Relevant Notes:
- [[agentic Taylorism means humanity feeds knowledge into AI through usage as a byproduct of labor and whether this concentrates or distributes depends entirely on engineering and evaluation]] — the concentration/distribution fork applies to AI-as-finance just as it does to AI-as-knowledge-extraction
- [[the price of anarchy quantifies the gap between cooperative optimum and competitive equilibrium and applying this framework to civilizational coordination failures offers a quantitative lens though operationalizing it at scale remains unproven]] — if AI can close the gap between competitive equilibrium and cooperative optimum directly, the PoA framework measures exactly what AI-finance would eliminate
Topics:
- [[_map]]

View file

@ -1,42 +1,39 @@
--- ---
type: claim type: claim
domain: mechanisms domain: mechanisms
description: "A pre-deployment assessment methodology that maps full affordance chains across environment, health, psychology, communities, power dynamics, and arms race potential — asking not just will this break but what else will this touch" description: "Cross-domain pre-deployment assessment that maps full affordance chains produces categorically different outcomes than domain-specific red teaming — social media's catastrophic effects were nth-order affordance cascades that no domain-specific assessment would have caught"
confidence: experimental confidence: experimental
source: "Schmachtenberger 'Development in Progress' (2024) Part II, extending military red team/blue team methodology" source: "Schmachtenberger 'Development in Progress' (2024) Part II, extending military red team/blue team methodology"
created: 2026-04-03 created: 2026-04-03
related: related:
- "for a change to equal progress it must systematically identify and internalize its externalities because immature progress that ignores cascading harms is the most dangerous ideology in the world" - "for a change to equal progress it must systematically identify and internalize its externalities because immature progress that ignores cascading harms is the most dangerous ideology in the world"
- "epistemic commons degradation is the gateway failure that enables all other civilizational risks because you cannot coordinate on problems you cannot collectively perceive"
--- ---
# Yellow teaming assesses all nth-order effects across domains before deployment distinct from red teaming which tests only for direct failure modes # Cross-domain pre-deployment assessment produces categorically different risk identification than domain-specific red teaming because the most catastrophic technology effects are nth-order affordance cascades invisible within any single domain
Schmachtenberger proposes yellow teaming as a distinct pre-deployment methodology that extends beyond red teaming's focus on adversarial failure modes. Where red teaming asks "can this be broken?" and blue teaming asks "can we defend it?", yellow teaming asks "what else will this touch?" Schmachtenberger proposes "yellow teaming" as a distinct pre-deployment methodology. Where red teaming asks "can this be broken?" and blue teaming asks "can we defend it?", yellow teaming asks "what else will this touch?" — mapping full affordance chains across environment, health, psychology, communities, power dynamics, and arms race potential.
The methodology requires mapping the full affordance chain of a technology or intervention: The arguable claim is not the methodology's existence but its necessity: **the most catastrophic effects of exponential technologies are nth-order cascades that cross domain boundaries and are therefore invisible to any domain-specific assessment.**
- **Environmental effects** — resource extraction, energy consumption, waste, ecosystem disruption
- **Health effects** — direct and indirect, acute and chronic, physical and psychological
- **Psychological effects** — attention, cognition, social comparison, addiction pathways
- **Community effects** — social cohesion, economic distribution, cultural disruption
- **Power dynamics** — concentration vs distribution of capability, surveillance potential, control affordances
- **Arms race potential** — whether the technology creates competitive dynamics that erode safety
The key distinction: red teaming is adversarial and domain-specific (does this AI system have jailbreaks?). Yellow teaming is systemic and cross-domain (if this AI system succeeds as designed, what secondary and tertiary effects cascade through social, economic, and political systems?). The social media case is the strongest evidence. Domain-specific red teaming would have caught privacy vulnerabilities, content moderation gaps, and platform stability issues. It would NOT have caught: the attention economy's effect on democratic sensemaking, adolescent mental health epidemics from social comparison algorithms, epistemic polarization from engagement optimization, or the weaponization of recommendation algorithms for political manipulation. These were not failure modes — they were success modes. The platform worked exactly as designed; the catastrophic effects were nth-order affordance cascades across psychology, politics, and epistemology.
Schmachtenberger's worked example is social media: red teaming would have caught privacy vulnerabilities and content moderation gaps. Yellow teaming would have caught the attention economy's effect on democratic sensemaking, adolescent mental health, epistemic polarization, and the weaponization of engagement algorithms for political manipulation. The direct failure modes were not the catastrophic ones — the nth-order affordance effects were. If this pattern generalizes — if exponential technologies consistently produce their worst effects through cross-domain cascades rather than direct failure — then domain-specific assessment is structurally inadequate for governing them. AI, synthetic biology, and neurotechnology all have cross-domain affordance profiles that suggest the same pattern.
The operational gap: yellow teaming has no institutional track record at scale. No company, government, or international body has implemented systematic cross-domain pre-deployment assessment for exponential technologies. The closest precedents are environmental impact assessments (narrow in scope) and technology assessment offices (historically defunded — the US Office of Technology Assessment was eliminated in 1995). **The operational gap is real:** No company, government, or international body has implemented systematic cross-domain pre-deployment assessment at scale. The closest precedents are environmental impact assessments (narrow in scope) and technology assessment offices (historically defunded — the US Office of Technology Assessment was eliminated in 1995). Whether yellow teaming is institutionally feasible or merely a good idea that can't be implemented under competitive pressure is the open question.
## Challenges ## Challenges
- Yellow teaming at full scope may be computationally and institutionally intractable. Mapping nth-order effects across all domains requires predictive capacity that may exceed what any team (human or AI-assisted) can achieve. - Yellow teaming at full scope may be computationally intractable. Mapping nth-order effects across all domains requires predictive capacity that may exceed what any team can achieve. The social media case is clear in hindsight; predicting AI's nth-order effects in advance may be qualitatively harder.
- The methodology risks analysis paralysis — if every technology must pass a full cross-domain assessment before deployment, innovation slows dramatically. The precautionary principle tension is real. - The methodology risks analysis paralysis. If every exponential technology must pass a full cross-domain assessment before deployment, innovation slows dramatically and competitive dynamics (Moloch) ensure non-compliant actors deploy first.
- Without institutional backing and enforcement, yellow teaming is advisory. Schmachtenberger provides no mechanism for ensuring yellow team findings are acted upon — the same competitive dynamics (Moloch) that produce externalities will pressure actors to ignore yellow team results. - Without enforcement mechanisms, yellow teaming is advisory. Schmachtenberger provides no mechanism for ensuring results are acted upon — the same competitive dynamics that produce externalities will pressure actors to ignore yellow team findings. The gap between identifying problems and creating incentives to address them is precisely the gap between Schmachtenberger's framework and mechanism design approaches.
- The social media case may not generalize. Social media's nth-order effects were severe because it directly modified human cognition and social behavior at scale. Not all exponential technologies have this profile — some may have effects that are catastrophic but domain-contained.
--- ---
Relevant Notes: Relevant Notes:
- [[for a change to equal progress it must systematically identify and internalize its externalities because immature progress that ignores cascading harms is the most dangerous ideology in the world]] — yellow teaming is the operational methodology for the progress redefinition - [[for a change to equal progress it must systematically identify and internalize its externalities because immature progress that ignores cascading harms is the most dangerous ideology in the world]] — yellow teaming is the operational methodology for the progress redefinition
- [[epistemic commons degradation is the gateway failure that enables all other civilizational risks because you cannot coordinate on problems you cannot collectively perceive]] — social media's effect on sensemaking is the paradigm case of nth-order affordance cascade
Topics: Topics:
- [[_map]] - [[_map]]