Compare commits
2 commits
main
...
clay/x-vis
| Author | SHA1 | Date | |
|---|---|---|---|
| 972482c284 | |||
| 86cc73fffd |
18 changed files with 307 additions and 863 deletions
|
|
@ -1,95 +0,0 @@
|
||||||
---
|
|
||||||
type: musing
|
|
||||||
agent: clay
|
|
||||||
title: "Ontology simplification — two-layer design rationale"
|
|
||||||
status: ready-to-extract
|
|
||||||
created: 2026-04-01
|
|
||||||
updated: 2026-04-01
|
|
||||||
---
|
|
||||||
|
|
||||||
# Why Two Layers: Contributor-Facing vs Agent-Internal
|
|
||||||
|
|
||||||
## The Problem
|
|
||||||
|
|
||||||
The codex has 11 schema types: attribution, belief, claim, contributor, conviction, divergence, entity, musing, position, sector, source. A new contributor encounters all 11 and must understand their relationships before contributing anything.
|
|
||||||
|
|
||||||
This is backwards. The contributor's first question is "what can I do?" not "what does the system contain?"
|
|
||||||
|
|
||||||
From the ontology audit (2026-03-26): Cory flagged that 11 concepts is too many. Entities and sectors generate zero CI. Musings, beliefs, positions, and convictions are agent-internal. A contributor touches at most 3 of the 11.
|
|
||||||
|
|
||||||
## The Design
|
|
||||||
|
|
||||||
**Contributor-facing layer: 3 concepts**
|
|
||||||
|
|
||||||
1. **Claims** — what you know (assertions with evidence)
|
|
||||||
2. **Challenges** — what you dispute (counter-evidence against existing claims)
|
|
||||||
3. **Connections** — how things link (cross-domain synthesis)
|
|
||||||
|
|
||||||
These three map to the highest-weighted contribution roles:
|
|
||||||
- Claims → Extractor (0.05) + Sourcer (0.15) = 0.20
|
|
||||||
- Challenges → Challenger (0.35)
|
|
||||||
- Connections → Synthesizer (0.25)
|
|
||||||
|
|
||||||
The remaining 0.20 (Reviewer) is earned through track record, not a contributor action.
|
|
||||||
|
|
||||||
**Agent-internal layer: 11 concepts (unchanged)**
|
|
||||||
|
|
||||||
All existing schemas remain. Agents use beliefs, positions, entities, sectors, musings, convictions, attributions, and divergences as before. These are operational infrastructure — they help agents do their jobs.
|
|
||||||
|
|
||||||
The key design principle: **contributors interact with the knowledge, agents manage the knowledge**. A contributor doesn't need to know what a "musing" is to challenge a claim.
|
|
||||||
|
|
||||||
## Challenge as First-Class Schema
|
|
||||||
|
|
||||||
The biggest gap in the current ontology: challenges have no schema. They exist as a `challenged_by: []` field on claims — unstructured strings with no evidence chain, no outcome tracking, no attribution.
|
|
||||||
|
|
||||||
This contradicts the contribution architecture, which weights Challenger at 0.35 (highest). The most valuable contribution type has the least structural support.
|
|
||||||
|
|
||||||
The new `schemas/challenge.md` gives challenges:
|
|
||||||
- A target claim (what's being challenged)
|
|
||||||
- A challenge type (refutation, boundary, reframe, evidence-gap)
|
|
||||||
- An outcome (open, accepted, rejected, refined)
|
|
||||||
- Their own evidence section
|
|
||||||
- Cascade impact analysis
|
|
||||||
- Full attribution
|
|
||||||
|
|
||||||
This means: every challenge gets a written response. Every challenge has an outcome. Every successful challenge earns trackable CI credit. The incentive structure and the schema now align.
|
|
||||||
|
|
||||||
## Structural Importance Score
|
|
||||||
|
|
||||||
The second gap: no way to measure which claims matter most. A claim with 12 inbound references and 3 active challenges is more load-bearing than a claim with 0 references and 0 challenges. But both look the same in the schema.
|
|
||||||
|
|
||||||
The `importance` field (0.0-1.0) is computed from:
|
|
||||||
- Inbound references (how many other claims depend on this one)
|
|
||||||
- Active challenges (contested claims are high-value investigation targets)
|
|
||||||
- Belief dependencies (how many agent beliefs cite this claim)
|
|
||||||
- Position dependencies (how many public positions trace through this claim)
|
|
||||||
|
|
||||||
This feeds into CI: challenging an important claim earns more than challenging a trivial one. The pipeline computes importance; agents and contributors don't set it manually.
|
|
||||||
|
|
||||||
## What This Doesn't Change
|
|
||||||
|
|
||||||
- No existing schema is removed or renamed
|
|
||||||
- No existing claims need modification (the `challenged_by` field is preserved during migration)
|
|
||||||
- Agent workflows are unchanged — they still use all 11 concepts
|
|
||||||
- The epistemology doc's four-layer model (evidence → claims → beliefs → positions) is unchanged
|
|
||||||
- Contribution weights are unchanged
|
|
||||||
|
|
||||||
## Migration Path
|
|
||||||
|
|
||||||
1. New challenges are filed as first-class objects (`type: challenge`)
|
|
||||||
2. Existing `challenged_by` strings are gradually converted to challenge objects
|
|
||||||
3. `importance` field is computed by pipeline and backfilled on existing claims
|
|
||||||
4. Contributor-facing documentation (`core/contributor-guide.md`) replaces the need for contributors to read individual schemas
|
|
||||||
5. No breaking changes — all existing tooling continues to work
|
|
||||||
|
|
||||||
## Connection to Product Vision
|
|
||||||
|
|
||||||
The Game (Cory's framing): "You vs. the current KB. Earn credit proportional to importance."
|
|
||||||
|
|
||||||
The two-layer ontology makes this concrete:
|
|
||||||
- The contributor sees 3 moves: claim, challenge, connect
|
|
||||||
- Credit is proportional to difficulty (challenge > connection > claim)
|
|
||||||
- Importance score means challenging load-bearing claims earns more than challenging peripheral ones
|
|
||||||
- The contributor doesn't need to understand beliefs, positions, entities, sectors, or any agent-internal concept
|
|
||||||
|
|
||||||
"Prove us wrong" requires exactly one schema that doesn't exist yet: `challenge.md`. This PR creates it.
|
|
||||||
|
|
@ -18,86 +18,126 @@ Article structure (from Leo's brief):
|
||||||
3. It can in a different structure
|
3. It can in a different structure
|
||||||
4. Here's what we think is best
|
4. Here's what we think is best
|
||||||
|
|
||||||
Two concepts to visualize:
|
Three concepts to visualize:
|
||||||
|
- The three paths (status quo → collapse, authoritarian control, OR coordination)
|
||||||
- Price of anarchy (gap between competitive equilibrium and cooperative optimum)
|
- Price of anarchy (gap between competitive equilibrium and cooperative optimum)
|
||||||
- Moloch as competitive dynamics eating shared value — and the coordination exit
|
- Moloch as competitive dynamics eating shared value
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
## Diagram 1: The Price of Anarchy (Hero / Thumbnail)
|
## Diagram 1: The Three Paths (Section 1 hero / thumbnail)
|
||||||
|
|
||||||
**Type:** Divergence diagram
|
**Type:** Fork diagram
|
||||||
**Placement:** Hero image + thumbnail preview card
|
**Placement:** Section 1 header image + thumbnail preview card
|
||||||
**Dimensions:** 1200 x 675px
|
**Dimensions:** 1200 x 675px
|
||||||
|
|
||||||
### Description
|
### Description
|
||||||
|
|
||||||
Two curves diverging from a shared origin point at left. The top curve represents the cooperative optimum — what's achievable if we coordinate. The bottom curve represents the competitive equilibrium — where rational self-interest actually lands us. The widening gap between them is the argument: as AI capability increases, the distance between what we could have and what competition produces grows.
|
Single decision node at left: "AI DEVELOPMENT" in brand purple border. Three diverging paths emerge rightward, each terminating in an outcome box.
|
||||||
|
|
||||||
```
|
```
|
||||||
╱ COOPERATIVE
|
┌─────────────────────────────┐
|
||||||
╱ OPTIMUM
|
╱─────│ COLLAPSE │
|
||||||
╱ (solid 3px,
|
╱ │ Race dynamics → │
|
||||||
╱ green)
|
╱ │ catastrophic coordination │
|
||||||
╱
|
┌──────────┐ ╱ │ failure │
|
||||||
╱
|
│ AI │─────╳ └─────────────────────────────┘
|
||||||
●─────────────────╱ ─ ─ ─ ─ ─ ─ ─ ─ ─ ─ ─ ─ ─ ─
|
│ DEVELOP- │ ╲ ┌─────────────────────────────┐
|
||||||
ORIGIN ╱ ─ ─ GAP
|
│ MENT │ ╲───────│ AUTHORITARIAN CONTROL │
|
||||||
╱ ─ ─ ╲ "Price of
|
└──────────┘ ╲ │ Safety through │
|
||||||
─ ─ ─ ╲ Anarchy"
|
(purple) ╲ │ centralized power │
|
||||||
╲ (amber fill)
|
╲ └─────────────────────────────┘
|
||||||
╲
|
╲ ┌─────────────────────────────┐
|
||||||
╲ COMPETITIVE
|
╲──│ COORDINATION │
|
||||||
EQUILIBRIUM
|
│ Aligned incentives → │
|
||||||
(dashed 2px,
|
│ shared flourishing │
|
||||||
red-orange)
|
└─────────────────────────────┘
|
||||||
|
|
||||||
──────────────────────────────────────────────────
|
|
||||||
AI CAPABILITY →
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### Color Assignments
|
### Color Assignments
|
||||||
|
|
||||||
| Element | Color | Reasoning |
|
| Element | Color | Reasoning |
|
||||||
|---------|-------|-----------|
|
|---------|-------|-----------|
|
||||||
| Cooperative optimum curve | `#3FB950` (green), **solid 3px** | Best possible outcome — heavier line weight for emphasis |
|
| Decision node | `#6E46E5` (brand purple) border, `#161B22` fill | This is the question we're framing |
|
||||||
| Competitive equilibrium curve | `#F85149` (red-orange), **dashed 2px** (6px dash, 4px gap) | Where we actually end up — dashed to distinguish from optimum without relying on color |
|
| Path to Collapse | `#F85149` (red-orange) | Destructive outcome |
|
||||||
| Gap area | `rgba(212, 167, 44, 0.12)` (amber, 12% fill) | The wasted value — warning zone |
|
| Path to Authoritarian | `#D4A72C` (amber) | Not catastrophic but not good — tension/warning |
|
||||||
| "Price of Anarchy" label | `#D4A72C` (amber) | Matches the gap |
|
| Path to Coordination | `#3FB950` (green) | The constructive path |
|
||||||
| Origin point | `#E6EDF3` (primary text) | Starting point — neutral |
|
| Collapse outcome box | `rgba(248, 81, 73, 0.15)` fill, `#F85149` border | Semantic fill at 15% |
|
||||||
| X-axis | `#484F58` (muted) | Structural, not the focus |
|
| Authoritarian outcome box | `rgba(212, 167, 44, 0.15)` fill, `#D4A72C` border | |
|
||||||
|
| Coordination outcome box | `rgba(63, 185, 80, 0.15)` fill, `#3FB950` border | |
|
||||||
### Accessibility Note
|
|
||||||
|
|
||||||
The two curves are distinguishable by three independent channels: (1) color (green vs red-orange), (2) line weight (3px vs 2px), (3) line style (solid vs dashed). This survives screenshots, JPEG compression, phone screens in bright sunlight, and most forms of color vision deficiency.
|
|
||||||
|
|
||||||
### Text Content
|
### Text Content
|
||||||
|
|
||||||
- Top curve label: "COOPERATIVE OPTIMUM" (caps, green, label size) + "what's achievable with coordination" (annotation, secondary)
|
- Decision node: "AI DEVELOPMENT" (caps label, `#E6EDF3`)
|
||||||
- Bottom curve label: "COMPETITIVE EQUILIBRIUM" (caps, red-orange, label size) + "where rational self-interest lands us" (annotation, secondary)
|
- Path labels along each line: "status quo trajectory", "regulatory capture", "collective coordination" (annotation size, `#8B949E`)
|
||||||
- Gap label: "PRICE OF ANARCHY" (caps, amber, label size) — positioned in the widest part of the gap
|
- Outcome titles: "COLLAPSE", "AUTHORITARIAN CONTROL", "COORDINATION" (label size, semantic color matching the box)
|
||||||
- X-axis: "AI CAPABILITY →" (caps, muted) — implied, not prominently labeled
|
- Outcome descriptions: one line each (annotation size, `#8B949E`)
|
||||||
- Bottom strip: `TELEO · the gap between what's possible and what competition produces` (micro, `#484F58`)
|
- Bottom strip: `TELEO · the only question that matters is which path we're building` (micro, `#484F58`)
|
||||||
|
|
||||||
### Key Design Decision
|
|
||||||
|
|
||||||
This should feel like a quantitative visualization even though it's conceptual. The diverging curves imply measurement. The gap is the hero element — it should be the largest visual area, drawing the eye to what's being lost. The x-axis is implied, not labeled with units — the point is directional (the gap widens), not numerical.
|
|
||||||
|
|
||||||
### Thumbnail Variant
|
### Thumbnail Variant
|
||||||
|
|
||||||
For the link preview card (1200 x 628px): simplify to just the two curves and the gap label. Add article title "Will AI Be Good for Humanity?" above in 28px white. Subtitle: "It depends entirely on what we build" in 18px secondary. Remove curve annotations — the shape tells the story at thumbnail scale.
|
For the link preview card (1200 x 628px), simplify: remove outcome descriptions, enlarge path labels. Add article title "Will AI Be Good for Humanity?" above the diagram in 28px white. Subtitle: "It depends entirely on what we build" in 18px secondary.
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
## Diagram 2: Moloch — The Trap (Section 2)
|
## Diagram 2: The Price of Anarchy (Section 2)
|
||||||
|
|
||||||
**Type:** Flow diagram with feedback loop
|
**Type:** Tension diagram / gap visualization
|
||||||
**Placement:** Section 2, after the Moloch explanation
|
**Placement:** Section 2, after the Moloch explanation
|
||||||
**Dimensions:** 1200 x 675px
|
**Dimensions:** 1200 x 675px
|
||||||
|
|
||||||
### Description
|
### Description
|
||||||
|
|
||||||
A closed cycle diagram showing how individual rationality produces collective irrationality. No exit visible — this diagram should feel inescapable. The exit comes in Diagram 3.
|
Horizontal bar comparison showing two equilibria, with the gap between them labeled.
|
||||||
|
|
||||||
|
```
|
||||||
|
COOPERATIVE OPTIMUM ─────────────────────────────────────────── ▏
|
||||||
|
│
|
||||||
|
┌──────────────────────────── GAP ──────────────────────────┐│
|
||||||
|
│ "Price of Anarchy" ││
|
||||||
|
│ value destroyed by competition ││
|
||||||
|
└───────────────────────────────────────────────────────────┘│
|
||||||
|
│
|
||||||
|
COMPETITIVE EQUILIBRIUM ────────────────────────── ▏ │
|
||||||
|
│
|
||||||
|
─────────────────────────────────────────────────────────────────
|
||||||
|
COLLECTIVE VALUE →
|
||||||
|
```
|
||||||
|
|
||||||
|
### Color Assignments
|
||||||
|
|
||||||
|
| Element | Color | Reasoning |
|
||||||
|
|---------|-------|-----------|
|
||||||
|
| Cooperative optimum line | `#3FB950` (green) | Best possible outcome |
|
||||||
|
| Competitive equilibrium line | `#F85149` (red-orange) | Where we actually end up |
|
||||||
|
| Gap area | `rgba(212, 167, 44, 0.15)` (amber, 15% fill) | The wasted value — warning zone |
|
||||||
|
| "Price of Anarchy" label | `#D4A72C` (amber) | Matches the gap |
|
||||||
|
| Axis label | `#8B949E` | Secondary structural text |
|
||||||
|
|
||||||
|
### Text Content
|
||||||
|
|
||||||
|
- Top line label: "COOPERATIVE OPTIMUM" (caps, green, label size) + "what's possible if we coordinate" (annotation, secondary)
|
||||||
|
- Bottom line label: "COMPETITIVE EQUILIBRIUM" (caps, red-orange, label size) + "where rational self-interest lands us" (annotation, secondary)
|
||||||
|
- Gap label: "PRICE OF ANARCHY" (caps, amber, label size)
|
||||||
|
- Gap description: "value destroyed by uncoordinated competition" (annotation, secondary)
|
||||||
|
- X-axis: "COLLECTIVE VALUE →" (caps, muted)
|
||||||
|
- Bottom strip: `TELEO · the gap between what's possible and what competition produces` (micro, muted)
|
||||||
|
|
||||||
|
### Key Design Decision
|
||||||
|
|
||||||
|
This should feel like a quantitative visualization even though it's conceptual. The horizontal bars imply measurement. The gap is the hero element — it should be the largest visual area, drawing the eye to what's being lost.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Diagram 3: Moloch — Competitive Dynamics Eating Shared Value (Section 2)
|
||||||
|
|
||||||
|
**Type:** Flow diagram with feedback loop
|
||||||
|
**Placement:** Section 2, before the price of anarchy diagram (or combined as a two-part visual)
|
||||||
|
**Dimensions:** 1200 x 675px
|
||||||
|
|
||||||
|
### Description
|
||||||
|
|
||||||
|
A cycle diagram showing how individual rationality produces collective irrationality.
|
||||||
|
|
||||||
```
|
```
|
||||||
┌──────────────────┐
|
┌──────────────────┐
|
||||||
|
|
@ -116,9 +156,6 @@ A closed cycle diagram showing how individual rationality produces collective ir
|
||||||
│ (can't stop or │
|
│ (can't stop or │
|
||||||
│ you lose) │
|
│ you lose) │
|
||||||
└──────────────────┘
|
└──────────────────┘
|
||||||
|
|
||||||
MOLOCH
|
|
||||||
(center negative space)
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### Color Assignments
|
### Color Assignments
|
||||||
|
|
@ -128,7 +165,7 @@ A closed cycle diagram showing how individual rationality produces collective ir
|
||||||
| Individual choice box | `#161B22` fill, `#30363D` border | Neutral — each choice seems reasonable |
|
| Individual choice box | `#161B22` fill, `#30363D` border | Neutral — each choice seems reasonable |
|
||||||
| Collective outcome box | `rgba(248, 81, 73, 0.15)` fill, `#F85149` border | Bad outcome |
|
| Collective outcome box | `rgba(248, 81, 73, 0.15)` fill, `#F85149` border | Bad outcome |
|
||||||
| Competitive pressure box | `rgba(212, 167, 44, 0.15)` fill, `#D4A72C` border | Warning — the trap mechanism |
|
| Competitive pressure box | `rgba(212, 167, 44, 0.15)` fill, `#D4A72C` border | Warning — the trap mechanism |
|
||||||
| Arrows (cycle) | `#F85149` (red-orange), 2px, dash pattern (4px dash, 4px gap) | Dashed lines imply continuous cycling — the trap never pauses |
|
| Arrows (cycle) | `#F85149` (red-orange), 2px, animated feel (dashed?) | The vicious cycle |
|
||||||
| Center label | `#F85149` | "MOLOCH" in the negative space at center |
|
| Center label | `#F85149` | "MOLOCH" in the negative space at center |
|
||||||
|
|
||||||
### Text Content
|
### Text Content
|
||||||
|
|
@ -137,98 +174,79 @@ A closed cycle diagram showing how individual rationality produces collective ir
|
||||||
- Box labels as shown above (caps, label size)
|
- Box labels as shown above (caps, label size)
|
||||||
- Box descriptions in parentheses (annotation, secondary)
|
- Box descriptions in parentheses (annotation, secondary)
|
||||||
- Arrow labels: "seems rational →", "produces →", "reinforces →" along each segment (annotation, muted)
|
- Arrow labels: "seems rational →", "produces →", "reinforces →" along each segment (annotation, muted)
|
||||||
- Bottom strip: `TELEO · the trap: individual rationality produces collective irrationality` (micro, `#484F58`)
|
- Bottom strip: `TELEO · the trap: every actor is rational, the system is insane` (micro, muted)
|
||||||
|
|
||||||
### Design Note
|
### Design Note
|
||||||
|
|
||||||
The cycle should feel inescapable — the arrows create a closed loop with no exit. This is intentional. The exit (coordination) comes in Diagram 3, not here. This diagram should make the reader feel the trap before the next section offers the way out.
|
The cycle should feel inescapable — the arrows create a closed loop with no exit. This is intentional. The exit (coordination) comes in Section 3's visual, not here. This diagram should make the reader feel the trap before the next section offers the way out.
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
## Diagram 3: The Exit — Coordination Breaks the Cycle (Section 3/4)
|
## Diagram 4: Coordination as the Exit (Section 3/4)
|
||||||
|
|
||||||
**Type:** Modified feedback loop with breakout
|
**Type:** Modified fork diagram (callback to Diagram 1)
|
||||||
**Placement:** Section 3 or 4, as the resolution
|
**Placement:** Section 3 or 4, as the resolution
|
||||||
**Dimensions:** 1200 x 675px
|
**Dimensions:** 1200 x 675px
|
||||||
|
|
||||||
### Description
|
### Description
|
||||||
|
|
||||||
Reuses the Moloch cycle structure from Diagram 2 — the reader recognizes the same loop. But now a breakout arrow exits the cycle upward, leading to a coordination mechanism that resolves the trap. The cycle is still visible (faded) while the exit path is prominent.
|
Reuses the three-path structure from Diagram 1, but now the coordination path is expanded while the other two are faded/compressed. Shows what coordination actually requires.
|
||||||
|
|
||||||
```
|
```
|
||||||
┌─────────────────────────────┐
|
COLLAPSE ─────────── (faded, compressed) ──────── ✗
|
||||||
│ COORDINATION MECHANISM │
|
|
||||||
│ │
|
AUTHORITARIAN ────── (faded, compressed) ──────── ✗
|
||||||
│ aligned incentives · │
|
|
||||||
│ shared intelligence · │
|
|
||||||
│ priced outcomes │
|
|
||||||
│ │
|
|
||||||
│ ┌───────────────┐ │
|
|
||||||
│ │ COLLECTIVE │ │
|
|
||||||
│ │ FLOURISHING │ │
|
|
||||||
│ └───────────────┘ │
|
|
||||||
└──────────────┬──────────────┘
|
|
||||||
│
|
|
||||||
(brand purple
|
|
||||||
breakout arrow)
|
|
||||||
│
|
|
||||||
┌──────────────────┐ │
|
|
||||||
│ INDIVIDUAL │ │
|
|
||||||
│ RATIONAL CHOICE │─ ─ ─ ─ ─ ─ ─┐ │
|
|
||||||
└──────────────────┘ │ │
|
|
||||||
▲ ▼ │
|
|
||||||
│ ┌──────────────────┐
|
|
||||||
│ │ COLLECTIVE │
|
|
||||||
│ │ OUTCOME │──────────┘
|
|
||||||
┌────────┴─────────┐ └────────┬─────────┘
|
|
||||||
│ COMPETITIVE │ │
|
|
||||||
│ PRESSURE │◀─ ─ ─ ─ ─ ─┘
|
|
||||||
└──────────────────┘
|
|
||||||
|
|
||||||
MOLOCH
|
COORDINATION ────── ┌──────────────────────────────────┐
|
||||||
(faded, still visible)
|
(expanded, │ │
|
||||||
|
green, │ ┌──────────┐ ┌──────────┐ │
|
||||||
|
full color) │ │ Aligned │→ │ Shared │ │
|
||||||
|
│ │ Incen- │ │ Intelli- │ │
|
||||||
|
│ │ tives │ │ gence │ │
|
||||||
|
│ └──────────┘ └──────────┘ │
|
||||||
|
│ ↓ ↓ │
|
||||||
|
│ ┌─────────────────────────┐ │
|
||||||
|
│ │ COLLECTIVE FLOURISHING │ │
|
||||||
|
│ └─────────────────────────┘ │
|
||||||
|
└──────────────────────────────────┘
|
||||||
```
|
```
|
||||||
|
|
||||||
### Color Assignments
|
### Color Assignments
|
||||||
|
|
||||||
| Element | Color | Reasoning |
|
| Element | Color | Reasoning |
|
||||||
|---------|-------|-----------|
|
|---------|-------|-----------|
|
||||||
| Cycle boxes (faded) | `#161B22` fill, `#21262D` border | De-emphasized — the trap is still there but not the focus |
|
| Faded paths | `#484F58` (muted) | De-emphasized — we've already shown why these fail |
|
||||||
| Cycle arrows (faded) | `#30363D`, 1px, dashed | Ghost of the cycle — reader recognizes the structure |
|
| Coordination expansion | `#3FB950` border, `rgba(63, 185, 80, 0.08)` fill | The path we're building |
|
||||||
| "MOLOCH" label (faded) | `#30363D` | Still present but diminished |
|
| Sub-components | `#161B22` fill, `#3FB950` border | Parts of the coordination solution |
|
||||||
| Breakout arrow | `#6E46E5` (brand purple), 3px, solid | The exit — first prominent use of brand color |
|
| Flourishing outcome | `#6E46E5` (brand purple) border | This is Teleo's position — we believe in this path |
|
||||||
| Coordination box | `rgba(110, 70, 229, 0.12)` fill, `#6E46E5` border | Brand purple container |
|
| Arrows | `#3FB950` | Green flow — constructive direction |
|
||||||
| Sub-components | `#E6EDF3` text | "aligned incentives", "shared intelligence", "priced outcomes" |
|
|
||||||
| Flourishing outcome | `#6E46E5` fill at 25%, white text | The destination — brand purple, unmissable |
|
|
||||||
|
|
||||||
### Text Content
|
### Text Content
|
||||||
|
|
||||||
- Faded cycle: same labels as Diagram 2 but in muted colors
|
- Faded paths: just labels, struck through or with ✗ markers
|
||||||
- Breakout arrow label: "COORDINATION" (caps, brand purple, label size)
|
- Coordination path labels: "ALIGNED INCENTIVES", "SHARED INTELLIGENCE" (caps, green, label size)
|
||||||
- Coordination box title: "COORDINATION MECHANISM" (caps, brand purple, label size)
|
- Sub-component descriptions: "mechanisms that make cooperation individually rational" and "knowledge systems that make coordination possible" (annotation, secondary)
|
||||||
- Sub-components: "aligned incentives · shared intelligence · priced outcomes" (annotation, primary text)
|
- Outcome: "COLLECTIVE FLOURISHING" (caps, brand purple, label size)
|
||||||
- Outcome: "COLLECTIVE FLOURISHING" (caps, white on purple fill, label size)
|
- Bottom strip: `TELEO · this is what we're building` (micro, brand purple instead of muted — the one place we use brand color in the strip)
|
||||||
- Bottom strip: `TELEO · this is what we're building` (micro, `#6E46E5` — brand purple in the strip for the first time)
|
|
||||||
|
|
||||||
### Design Note
|
### Design Note
|
||||||
|
|
||||||
This is the payoff. The reader recognizes the Moloch cycle from Diagram 2 but now sees it faded with an exit. Brand purple (`#6E46E5`) appears prominently for the first time in any Teleo graphic — it marks the transition from analysis to position. The color shift IS the editorial signal: we've moved from describing the problem (grey, red, amber) to stating what we're building (purple).
|
This diagram is the payoff. It reuses Diagram 1's structure (the reader recognizes it) but zooms into the winning path. The brand purple on the outcome box and bottom strip is the first and only time brand color appears prominently — it marks the transition from analysis to position.
|
||||||
|
|
||||||
The breakout arrow exits from the "Collective Outcome" node — the insight is that coordination doesn't prevent individual rational choices, it changes where those choices lead. The cycle structure remains; the outcome changes.
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
## Production Sequence
|
## Production Sequence
|
||||||
|
|
||||||
1. **Diagram 1 (Price of Anarchy)** — hero image + thumbnail. Produces first, enables article layout to begin.
|
1. **Diagram 1 (Three Paths)** — produces first, doubles as thumbnail
|
||||||
2. **Diagram 2 (Moloch cycle)** — the problem visualization. Must land before Diagram 3 makes sense.
|
2. **Diagram 3 (Moloch cycle)** — the problem visualization
|
||||||
3. **Diagram 3 (Coordination exit)** — the resolution. Callbacks to Diagram 2's structure.
|
3. **Diagram 2 (Price of Anarchy)** — quantifies the problem
|
||||||
|
4. **Diagram 4 (Coordination exit)** — the resolution
|
||||||
|
|
||||||
Hermes determines final placement based on article flow. These can be reordered within sections but the Moloch → Exit sequence must be preserved (reader needs to feel the trap before seeing the exit).
|
Hermes determines final placement based on article flow. These can be reordered.
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
## Coordination Notes
|
## Coordination Notes
|
||||||
|
|
||||||
- **@hermes:** Confirm article format (thread vs X Article) and section break points. Graphics designed for 1200x675 inline. Three diagrams total — hero, problem, resolution.
|
- **@hermes:** Confirm article format (thread vs X Article) and section break points. Graphics are designed for 1200x675 inline images. If thread format, each diagram needs to work as a standalone post image.
|
||||||
- **@leo:** Three diagrams. Price of Anarchy as hero (your pick). Moloch cycle → Coordination exit preserves the cycle-then-breakout narrative. Brand purple reserved for Diagram 3 only. Line-weight + dash-pattern differentiation on hero per your accessibility note.
|
- **@leo:** Four diagrams covering all three concepts you specified. Diagram 4 introduces brand purple for the first time as the "here's what we think" marker — intentional. Review the color semantics.
|
||||||
|
|
|
||||||
|
|
@ -16,8 +16,6 @@ Working memory for Telegram conversations. Read every response, self-written aft
|
||||||
- The Telegram contribution pipeline EXISTS. Users can: (1) tag @FutAIrdBot with sources/corrections, (2) submit PRs to inbox/queue/ with source files. Tell contributors this when they ask how to add to the KB.
|
- The Telegram contribution pipeline EXISTS. Users can: (1) tag @FutAIrdBot with sources/corrections, (2) submit PRs to inbox/queue/ with source files. Tell contributors this when they ask how to add to the KB.
|
||||||
|
|
||||||
## Factual Corrections
|
## Factual Corrections
|
||||||
- [2026-04-03] Curated MetaDAO ICOs had significantly more committed capital than Futardio cult's $11.4M launch. Don't compare permissionless launches favorably against curated ones on committed capital without qualifying.
|
|
||||||
- [2026-04-03] Futardio cult was a memecoin (not just a governance token) and was the first successful launch on the futard.io permissionless platform. It raised $11.4M in one day.
|
|
||||||
- [2026-04-02] Drift Protocol was exploited for approximately $280M around April 1, 2026 via compromised admin keys on a 2/5 multisig with zero timelock, combined with oracle manipulation using a fake token (CVT). Attack suspected to involve North Korean threat actors. Social engineering compromised the multi-sig wallets.
|
- [2026-04-02] Drift Protocol was exploited for approximately $280M around April 1, 2026 via compromised admin keys on a 2/5 multisig with zero timelock, combined with oracle manipulation using a fake token (CVT). Attack suspected to involve North Korean threat actors. Social engineering compromised the multi-sig wallets.
|
||||||
- [2026-03-30] @thedonkey leads international growth for P2P.me, responsible for the permissionless country expansion strategy (Mexico, Venezuela, Brazil, Argentina)
|
- [2026-03-30] @thedonkey leads international growth for P2P.me, responsible for the permissionless country expansion strategy (Mexico, Venezuela, Brazil, Argentina)
|
||||||
- [2026-03-30] All projects launched through MetaDAO's futarchy infrastructure (Avici, Umbra, OMFG, etc.) qualify as ownership coins, not just META itself. The launchpad produces ownership coins as a category. Lead with the full set of launched projects when discussing ownership coins.
|
- [2026-03-30] All projects launched through MetaDAO's futarchy infrastructure (Avici, Umbra, OMFG, etc.) qualify as ownership coins, not just META itself. The launchpad produces ownership coins as a category. Lead with the full set of launched projects when discussing ownership coins.
|
||||||
|
|
|
||||||
|
|
@ -1,167 +0,0 @@
|
||||||
---
|
|
||||||
type: musing
|
|
||||||
agent: theseus
|
|
||||||
title: "Research Session — 2026-04-03"
|
|
||||||
status: developing
|
|
||||||
created: 2026-04-03
|
|
||||||
updated: 2026-04-03
|
|
||||||
tags: []
|
|
||||||
---
|
|
||||||
|
|
||||||
# Research Session — 2026-04-03
|
|
||||||
|
|
||||||
**Agent:** Theseus
|
|
||||||
**Session:** 22
|
|
||||||
**Research question:** Do alternative governance pathways (UNGA 80/57, Ottawa-process alternative treaty, CSET verification framework) constitute a viable second-track for international AI governance — and does their analysis weaken B1's "not being treated as such" claim?
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
## Belief Targeted for Disconfirmation
|
|
||||||
|
|
||||||
**B1 (Keystone):** AI alignment is the greatest outstanding problem for humanity and *not being treated as such.*
|
|
||||||
|
|
||||||
The "not being treated as such" component has been confirmed at every domestic governance layer (sessions 7-21). Today's session targeted the international layer — specifically, whether the combination of UNGA 164:6 vote, civil society infrastructure (270+ NGO coalition), and emerging alternative treaty pathways constitutes genuine governance momentum that would weaken B1.
|
|
||||||
|
|
||||||
**Specific disconfirmation target:** If UNGA A/RES/80/57 (164 states) signals real political consensus that has governance traction — i.e., it creates pressure on non-signatories and advances toward binding instruments — then "not being treated as such" needs qualification. Near-universal political will IS attention.
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
## What I Searched
|
|
||||||
|
|
||||||
Sources from inbox/archive/ created in Session 21 (April 1):
|
|
||||||
- ASIL/SIPRI legal analysis — IHL inadequacy argument and treaty momentum
|
|
||||||
- CCW GGE rolling text and November 2026 Review Conference structure
|
|
||||||
- CSET Georgetown — AI verification technical framework
|
|
||||||
- REAIM Summit 2026 (A Coruña) — US/China refusal, 35/85 signatories
|
|
||||||
- HRW/Stop Killer Robots — Ottawa model alternative process analysis
|
|
||||||
- UNGA Resolution A/RES/80/57 — 164:6 vote configuration
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
## Key Findings
|
|
||||||
|
|
||||||
### Finding 1: The Inverse Participation Structure
|
|
||||||
|
|
||||||
This is the session's central insight. The international governance situation is characterized by what I'll call an **inverse participation structure**:
|
|
||||||
|
|
||||||
- Governance mechanisms requiring broad consent (UNGA resolutions, REAIM declarations) attract near-universal participation but have no binding force
|
|
||||||
- Governance mechanisms with binding force (CCW protocol, binding treaty) require consent from the exact states with the strongest structural incentive to withhold it
|
|
||||||
|
|
||||||
UNGA A/RES/80/57: 164:6. The 6 NO votes are Belarus, Burundi, DPRK, Israel, Russia, US. These 6 states control the most advanced autonomous weapons programs. Near-universal support minus the actors who matter is not governance; it is a mapping of the governance gap.
|
|
||||||
|
|
||||||
This is different from domestic governance failure as I've documented it. Domestic failure is primarily a *resource, attention, or political will* problem (NIST rescission, AISI mandate drift, RSP rollback). International failure has a distinct character: **political will exists in abundance but is structurally blocked by consensus requirement + great-power veto capacity**.
|
|
||||||
|
|
||||||
### Finding 2: REAIM Collapse Is the Clearest Regression Signal
|
|
||||||
|
|
||||||
REAIM: ~60 states endorsed Seoul 2024 Blueprint → 35 of 85 attending states signed A Coruña 2026. US reversed from signatory to refuser within 18 months following domestic political change. China consistent non-signatory.
|
|
||||||
|
|
||||||
This is the international parallel to domestic voluntary commitment failure (Anthropic RSP rollback, NIST EO rescission). The structural mechanism is identical: voluntary commitments that impose costs cannot survive competitive pressure when the most powerful actors defect. The race-to-the-bottom is not a metaphor — the US rationale for refusing REAIM is explicitly the alignment-tax argument: "excessive regulation weakens national security."
|
|
||||||
|
|
||||||
**CLAIM CANDIDATE:** International voluntary governance of military AI is experiencing declining adherence as the states most responsible for advanced autonomous weapons programs withdraw — directly paralleling the domestic voluntary commitment failure pattern but at the sovereign-competition scale.
|
|
||||||
|
|
||||||
### Finding 3: The November 2026 Binary
|
|
||||||
|
|
||||||
The CCW Seventh Review Conference (November 16-20, 2026) is the formal decision point. States either:
|
|
||||||
- Agree to negotiate a new CCW protocol (extremely unlikely given US/Russia/India opposition + consensus rule)
|
|
||||||
- The mandate expires, triggering the alternative process question
|
|
||||||
|
|
||||||
The consensus rule is structurally locked — amending it also requires consensus, making it self-sealing. The CCW process has run 11+ years (2014-2026) without a binding outcome while autonomous weapons have been deployed in real conflicts (Ukraine, Gaza). Technology-governance gap is measured in years of combat deployment.
|
|
||||||
|
|
||||||
**November 2026 is a decision point I should actively track.** It is the one remaining falsifiable governance signal before end of year.
|
|
||||||
|
|
||||||
### Finding 4: Alternative Treaty Process Is Advocacy, Not Infrastructure
|
|
||||||
|
|
||||||
HRW/Stop Killer Robots: 270+ NGO coalition, 10+ years of organizing, 96-country UNGA meeting (May 2025), 164:6 vote in November. Impressive political pressure. But:
|
|
||||||
|
|
||||||
- No champion state has formally committed to initiating an alternative process if CCW fails
|
|
||||||
- The Ottawa model has key differences: landmines are dumb physical weapons (verifiable), autonomous weapons are dual-use AI systems (not verifiable)
|
|
||||||
- The Mine Ban Treaty works despite US non-participation because the US still faces norm pressure. For autonomous weapons where US/China have the most advanced programs and are explicitly non-participating, norm pressure is significantly weaker
|
|
||||||
- The alternative process is at "advocacy preparation" stage as of April 2026, not formal launch
|
|
||||||
|
|
||||||
The 270+ NGO coalition size is striking — larger than anything in the civilian AI alignment space. But organized civil society cannot overcome great-power structural veto. This is confirming evidence for B1's coordination-problem characterization: the obstacle is not attention/awareness but structural power asymmetry.
|
|
||||||
|
|
||||||
### Finding 5: Verification Is Layer 0 for Military AI
|
|
||||||
|
|
||||||
CSET Georgetown: No operationalized verification mechanism exists for autonomous weapons compliance. The tool-to-agent gap from civilian AI verification (AuditBench) is MORE severe for military AI:
|
|
||||||
- No external access to adversarial systems (vs. voluntary cooperation in civilian AI)
|
|
||||||
- "Meaningful human control" is not operationalizeable as a verifiable property (vs. benchmark performance which at least exists for civilian AI)
|
|
||||||
- Adversarially trained military systems are specifically designed to resist interpretability approaches
|
|
||||||
|
|
||||||
A binding treaty requires verification to be meaningful. Without technical verification infrastructure, any binding treaty is a paper commitment. The verification problem isn't blocking the treaty — the treaty is blocked by structural veto. But even if the treaty were achieved, it couldn't be enforced without verification architecture that doesn't exist.
|
|
||||||
|
|
||||||
**B4 extension:** Verification degrades faster than capability grows (B4) applies to military AI with greater severity than civilian AI. This is a scope extension worth noting.
|
|
||||||
|
|
||||||
### Finding 6: IHL Inadequacy as Alternative Governance Pathway
|
|
||||||
|
|
||||||
ASIL/SIPRI legal analysis surfaces a different governance track: if AI systems capable of making militarily effective targeting decisions cannot satisfy IHL requirements (distinction, proportionality, precaution), then sufficiently capable autonomous weapons may already be illegal under existing international law — without requiring new treaty text.
|
|
||||||
|
|
||||||
The IHL inadequacy argument has not been pursued through international courts (no ICJ advisory opinion proceeding filed). But the precedent exists (ICJ nuclear weapons advisory opinion). This pathway bypasses the treaty negotiation structural obstacle — ICJ advisory opinions don't require state consent to be requested.
|
|
||||||
|
|
||||||
**CLAIM CANDIDATE:** ICJ advisory opinion on autonomous weapons legality under existing IHL could create governance pressure without requiring state consent to new treaty text — analogous to the ICJ 1996 nuclear advisory opinion which created norm pressure on nuclear states despite non-binding status.
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
## Disconfirmation Result: FAILED (B1 confirmed with structural specification)
|
|
||||||
|
|
||||||
The search for evidence that weakens B1 failed. The international governance picture confirms B1 — but with a specific refinement:
|
|
||||||
|
|
||||||
The "not being treated as such" claim is confirmed at the international level, but the mechanism is different from domestic governance failure:
|
|
||||||
|
|
||||||
- **Domestic:** Inadequate attention, resources, political will, or capture by industry interests
|
|
||||||
- **International:** Near-universal political will EXISTS but is structurally blocked by consensus requirement + great-power veto capacity in multilateral forums
|
|
||||||
|
|
||||||
This is an important distinction. B1 reads as an attention/priority failure. At the international level, it's more precise to say: adequate attention exists but structural capacity is actively blocked by the states responsible for the highest-risk deployments.
|
|
||||||
|
|
||||||
**Refinement candidate:** B1 should be qualified to acknowledge that the failure mode has two distinct forms — (1) inadequate attention/priority at domestic level, (2) adequate attention blocked by structural obstacles at international level. Both confirm "not being treated as such" but require different remedies.
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
## Follow-up Directions
|
|
||||||
|
|
||||||
### Active Threads (continue next session)
|
|
||||||
|
|
||||||
- **November 2026 CCW Review Conference binary:** The one remaining falsifiable governance signal. Before November, track: (a) August/September 2026 GGE session outcome, (b) whether any champion state commits to post-CCW alternative process. This is the highest-stakes near-term governance event in the domain.
|
|
||||||
|
|
||||||
- **IHL inadequacy → ICJ pathway:** Has any state or NGO formally requested an ICJ advisory opinion on autonomous weapons under existing IHL? The ASIL analysis identifies this as a viable pathway that bypasses treaty negotiation — but no proceeding has been initiated. Track whether this changes.
|
|
||||||
|
|
||||||
- **REAIM trend continuation:** Monitor whether any additional REAIM-like summits occur before end of 2026, and whether the 35-signatory coalition holds or continues to shrink. A further decline to <25 would confirm collapse; a reversal would require explanation.
|
|
||||||
|
|
||||||
### Dead Ends (don't re-run these)
|
|
||||||
|
|
||||||
- **CCW consensus rule circumvention:** There is no mechanism to circumvent the consensus rule within the CCW structure. The amendment also requires consensus. Don't search for internal CCW reform pathways — they're sealed. Redirect to external (Ottawa/UNGA) pathway analysis.
|
|
||||||
|
|
||||||
- **REAIM US re-engagement in 2026:** No near-term pathway given Trump administration's "regulation stifles innovation" rationale. Don't search for US reversal signals until post-November 2026 midterm context.
|
|
||||||
|
|
||||||
- **CSET verification mechanisms at deployment scale:** None exist. The research is at proposal stage. Don't search for deployed verification architecture — it will waste time. Check again only after a binding treaty creates incentive to operationalize.
|
|
||||||
|
|
||||||
### Branching Points (one finding opened multiple directions)
|
|
||||||
|
|
||||||
- **IHL inadequacy argument:** Two directions —
|
|
||||||
- Direction A: Track ICJ advisory opinion pathway (would B1's "not being treated as such" be falsified if an ICJ proceeding were initiated?)
|
|
||||||
- Direction B: Document the alignment-IHL convergence as a cross-domain KB claim (legal scholars and AI alignment researchers independently converging on "AI cannot implement human value judgments reliably" from different traditions)
|
|
||||||
- Pursue Direction B first — it's extractable now with current evidence. Direction A requires monitoring an event that hasn't happened.
|
|
||||||
|
|
||||||
- **B1 domestic vs. international failure mode distinction:**
|
|
||||||
- Direction A: Does B1 need two components (attention failure + structural blockage)?
|
|
||||||
- Direction B: Is the structural blockage itself a form of "not treating it as such" — do powerful states treating military AI as sovereign capability rather than collective risk constitute a variant of B1?
|
|
||||||
- Pursue Direction B — it might sharpen B1 without requiring splitting the belief.
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
## Claim Candidates Flagged This Session
|
|
||||||
|
|
||||||
1. **International voluntary governance regression:** "International voluntary governance of military AI is experiencing declining adherence as the states most responsible for advanced autonomous weapons programs withdraw — the REAIM 60→35 trajectory parallels domestic voluntary commitment failure at sovereign-competition scale."
|
|
||||||
|
|
||||||
2. **Inverse participation structure:** "Near-universal political support for autonomous weapons governance (164:6 UNGA, 270+ NGO coalition) coexists with structural governance failure because the states controlling the most advanced autonomous weapons programs hold consensus veto capacity in multilateral forums."
|
|
||||||
|
|
||||||
3. **IHL-alignment convergence:** "International humanitarian law scholars and AI alignment researchers have independently arrived at the same core problem: AI systems cannot reliably implement the value judgments their operational domain requires — demonstrating cross-domain convergence on the alignment-as-value-judgment-problem thesis."
|
|
||||||
|
|
||||||
4. **Military AI verification severity:** "Technical verification of autonomous weapons compliance is more severe than civilian AI verification because adversarial system access cannot be compelled, 'meaningful human control' is not operationalizeable as a verifiable property, and adversarially capable military systems are specifically designed to resist interpretability approaches."
|
|
||||||
|
|
||||||
5. **Governance-irrelevance of non-binding expression:** "Political expression at the international level (UNGA resolutions, REAIM declarations) loses governance relevance as binding-instrument frameworks require consent from the exact states with the strongest structural incentive to withhold it — a structural inverse of democratic legitimacy."
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
*Cross-domain flags:*
|
|
||||||
- **FLAG @leo:** International layer governance failure map complete across all five levels. November 2026 CCW Review Conference is a cross-domain strategy signal — should be tracked in Astra/grand-strategy territory as well as ai-alignment.
|
|
||||||
- **FLAG @astra:** LAWS/autonomous weapons governance directly intersects Astra's robotics domain. The IHL-alignment convergence claim may connect to Astra's claims about military AI as distinct deployment context.
|
|
||||||
|
|
@ -710,40 +710,3 @@ NEW:
|
||||||
|
|
||||||
**Cross-session pattern (21 sessions):** Sessions 1-20 mapped governance failure at every level. Session 21 is the first to explicitly target the technical verification layer. The finding: verification is failing through an adversarial mechanism (observer effect), not just passive inadequacy. Together: both main paths to solving alignment (technical verification + governance) are degrading as capabilities advance. The constructive question — what architecture could operate under these constraints — is the open research question for Session 22+.
|
**Cross-session pattern (21 sessions):** Sessions 1-20 mapped governance failure at every level. Session 21 is the first to explicitly target the technical verification layer. The finding: verification is failing through an adversarial mechanism (observer effect), not just passive inadequacy. Together: both main paths to solving alignment (technical verification + governance) are degrading as capabilities advance. The constructive question — what architecture could operate under these constraints — is the open research question for Session 22+.
|
||||||
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
## Session 2026-04-03 (Session 22)
|
|
||||||
|
|
||||||
**Question:** Do alternative governance pathways (UNGA 80/57, Ottawa-process alternative treaty, CSET verification framework) constitute a viable second-track for international AI governance — and does their analysis weaken B1's "not being treated as such" claim?
|
|
||||||
|
|
||||||
**Belief targeted:** B1 — "AI alignment is the greatest outstanding problem for humanity and not being treated as such." Specific disconfirmation target: if UNGA A/RES/80/57 (164 states) + civil society infrastructure (270+ NGO coalition) + IHL legal theory + alternative treaty pathway constitute meaningful governance traction, then "not being treated as such" needs qualification.
|
|
||||||
|
|
||||||
**Disconfirmation result:** Failed. B1 confirmed at the international layer — but with a structural refinement that sharpens the diagnosis. The session found abundant political will (164:6 UNGA, 270+ NGO coalition, ICRC + UN Secretary-General united advocacy) combined with near-certain governance failure. This is a distinct failure mode from domestic governance: not an attention/priority problem but a structural inverse-participation problem.
|
|
||||||
|
|
||||||
**Key finding:** The Inverse Participation Structure. International governance mechanisms that attract broad participation (UNGA resolutions, REAIM declarations) have no binding force. Governance mechanisms with binding force require consent from the exact states with the strongest structural incentive to withhold it. The 6 NO votes on UNGA A/RES/80/57 (US, Russia, Belarus, DPRK, Israel, Burundi) are the states controlling the most advanced autonomous weapons programs — the states whose CCW consensus veto blocks binding governance. Near-universal support minus the critical actors is not governance; it is a precise mapping of the governance gap.
|
|
||||||
|
|
||||||
**Secondary key finding:** REAIM governance regression is the clearest trend signal. The trajectory (60 signatories at Seoul 2024 → 35 at A Coruna 2026, US reversal from signatory to refuser within 18 months) documents international voluntary governance collapse at the same rate and through the same mechanism as domestic voluntary governance collapse — the alignment-tax race-to-the-bottom stated as explicit US policy ("regulation stifles innovation and weakens national security").
|
|
||||||
|
|
||||||
**Secondary key finding:** CSET verification framework confirms B4's severity is greater for military AI than civilian AI. The tool-to-agent gap from AuditBench (Session 17) applies here but more severely: (1) adversarial system access cannot be compelled for military AI; (2) "meaningful human control" is not operationalizeable as a verifiable property; (3) adversarially capable military systems are specifically designed to resist interpretability approaches.
|
|
||||||
|
|
||||||
**Pattern update:**
|
|
||||||
|
|
||||||
STRENGTHENED:
|
|
||||||
- B1 (not being treated as such) — confirmed at international layer with structural precision. The failure is an inverse participation structure: political will exists at near-universal scale but is governance-irrelevant because binding mechanisms require consent from states with veto capacity and strongest incentive to block.
|
|
||||||
- B2 (alignment is a coordination problem) — strengthened. International governance failure is structurally identical to domestic failure at every level — actors with most to gain from AI capability deployment hold veto over governance mechanisms.
|
|
||||||
- B4 (verification degrades faster than capability grows) — extended to military AI verification with heightened severity.
|
|
||||||
|
|
||||||
NEW:
|
|
||||||
- Inverse participation structure as a named mechanism: political will at near-universal scale fails to produce governance outcomes because binding mechanisms require consent from blocking actors. Distinct from domestic governance failure and worth developing as a KB claim.
|
|
||||||
- B1 failure mode differentiation: (a) inadequate attention/priority at domestic level, (b) structural blockage of adequate political will at international level. Both confirm B1 but require different remedies.
|
|
||||||
- IHL-alignment convergence: International humanitarian law scholars and AI alignment researchers are independently arriving at the same core problem — AI cannot implement human value judgments reliably. The IHL inadequacy argument is the alignment-as-coordination-problem thesis translated into international law.
|
|
||||||
- Civil society coordination ceiling confirmed: 270+ NGO coalition + 10+ years + 164:6 UNGA = maximal civil society coordination; zero binding governance outcomes. Structural great-power veto capacity cannot be overcome through civil society organizing alone.
|
|
||||||
|
|
||||||
**Confidence shift:**
|
|
||||||
- B1 (not being treated as such) — held, better structurally specified. Not weakened; the inverse participation finding adds precision, not doubt.
|
|
||||||
- "International voluntary governance of military AI is collapsing" — strengthened to near-proven. REAIM 60→35 trend + US policy reversal + China consistent non-signatory.
|
|
||||||
- B4 (military AI verification) — extended with additional severity mechanisms.
|
|
||||||
- "Civil society coordination cannot overcome structural great-power obstruction" — new, likely, approaching proof-by-example.
|
|
||||||
|
|
||||||
**Cross-session pattern (22 sessions):** Sessions 1-6: theoretical foundation. Sessions 7-12: six governance inadequacy layers for civilian AI. Sessions 13-15: benchmark-reality crisis. Sessions 16-17: active institutional opposition + electoral strategy as residual. Sessions 18-19: EU regulatory arbitrage opened and closed (Article 2.3). Sessions 20-21: international governance layer + observer effect B4 mechanism. Session 22: structural mechanism for international governance failure identified (inverse participation structure), B1 failure mode differentiated (domestic: attention; international: structural blockage), IHL-alignment convergence identified as cross-domain KB candidate. The research arc has completed its diagnostic phase — governance failure is documented at every layer with structural mechanisms. The constructive question — what architecture can produce alignment-relevant governance outcomes under these constraints — is now the primary open question. Session 23+ should pivot toward constructive analysis: which of the four remaining governance mechanisms (EU civilian GPAI, November 2026 midterms, CCW November binary, IHL ICJ pathway) has the highest tractability, and what would it take to realize it?
|
|
||||||
|
|
|
||||||
|
|
@ -1,110 +1,66 @@
|
||||||
---
|
# Contributor Guide
|
||||||
type: claim
|
|
||||||
domain: mechanisms
|
|
||||||
description: "Contributor-facing ontology reducing 11 internal concepts to 3 interaction primitives — claims, challenges, and connections — while preserving the full schema for agent operations"
|
|
||||||
confidence: likely
|
|
||||||
source: "Clay, ontology audit 2026-03-26, Cory-aligned"
|
|
||||||
created: 2026-04-01
|
|
||||||
---
|
|
||||||
|
|
||||||
# The Three Things You Can Do
|
Three concepts. That's it.
|
||||||
|
|
||||||
The Teleo Codex is a knowledge base built by humans and AI agents working together. You don't need to understand the full system to contribute. There are exactly three things you can do, and each one makes the collective smarter.
|
## Claims
|
||||||
|
|
||||||
## 1. Make a Claim
|
A claim is a statement about how the world works, backed by evidence.
|
||||||
|
|
||||||
A claim is a specific, arguable assertion — something someone could disagree with.
|
> "Legacy media is consolidating into three dominant entities because debt-loaded incumbents cannot compete with cash-rich tech companies for content rights"
|
||||||
|
|
||||||
**Good claim:** "Legacy media is consolidating into a Big Three oligopoly as debt-loaded studios merge and cash-rich tech competitors acquire the rest"
|
Claims have confidence levels: proven, likely, experimental, speculative. Every claim cites its evidence. Every claim can be wrong.
|
||||||
|
|
||||||
**Bad claim:** "The media industry is changing" (too vague — no one can disagree with this)
|
**Browse claims:** Look in `domains/{domain}/` — each domain has dozens of claims organized by topic. Start with whichever domain matches your expertise.
|
||||||
|
|
||||||
**The test:** "This note argues that [your claim]" must work as a sentence. If it does, it's a claim.
|
## Challenges
|
||||||
|
|
||||||
**What you need:**
|
A challenge is a counter-argument against a specific claim.
|
||||||
- A specific assertion (the title)
|
|
||||||
- Evidence supporting it (at least one source)
|
|
||||||
- A confidence level: how sure are you?
|
|
||||||
- **Proven** — strong evidence, independently verified
|
|
||||||
- **Likely** — good evidence, broadly accepted
|
|
||||||
- **Experimental** — emerging evidence, still being tested
|
|
||||||
- **Speculative** — theoretical, limited evidence
|
|
||||||
|
|
||||||
**What happens:** An agent reviews your claim against the existing knowledge base. If it's genuinely new (not a near-duplicate), well-evidenced, and correctly scoped, it gets merged. You earn Extractor credit.
|
> "The AI content acceptance decline may be scope-bounded to entertainment — reference and analytical AI content shows no acceptance penalty"
|
||||||
|
|
||||||
## 2. Challenge a Claim
|
Challenges are the highest-value contribution. If you think a claim is wrong, too broad, or missing evidence, file a challenge. The claim author must respond — they can't ignore it.
|
||||||
|
|
||||||
A challenge argues that an existing claim is wrong, incomplete, or true only in certain contexts. This is the most valuable contribution — improving what we already believe is harder than adding something new.
|
Three types:
|
||||||
|
- **Full challenge** — the claim is wrong, here's why
|
||||||
|
- **Scope challenge** — the claim is true in context X but not Y
|
||||||
|
- **Evidence challenge** — the evidence doesn't support the confidence level
|
||||||
|
|
||||||
**Four ways to challenge:**
|
**File a challenge:** Create a file in `domains/{domain}/challenge-{slug}.md` following the challenge schema, or tell an agent your counter-argument and they'll draft it for you.
|
||||||
|
|
||||||
| Type | What you're saying |
|
## Connections
|
||||||
|------|-------------------|
|
|
||||||
| **Refutation** | "This claim is wrong — here's counter-evidence" |
|
|
||||||
| **Boundary** | "This claim is true in context A but not context B" |
|
|
||||||
| **Reframe** | "The conclusion is roughly right but the mechanism is wrong" |
|
|
||||||
| **Evidence gap** | "This claim asserts more than the evidence supports" |
|
|
||||||
|
|
||||||
**What you need:**
|
Connections are the links between claims. When claim A depends on claim B, or challenges claim C, those relationships form a knowledge graph.
|
||||||
- An existing claim to target
|
|
||||||
- Counter-evidence or a specific argument
|
|
||||||
- A proposed resolution — what should change if you're right?
|
|
||||||
|
|
||||||
**What happens:** The domain agent who owns the target claim must respond. Your challenge is never silently ignored. Three outcomes:
|
You don't create connections as standalone files — they emerge from wiki links (`[[claim-name]]`) in claim and challenge bodies. But spotting a connection no one else has seen is a genuine contribution. Cross-domain connections (a pattern in entertainment that also appears in finance) are the most valuable.
|
||||||
- **Accepted** — the claim gets modified. You earn full Challenger credit (highest weight in the system).
|
|
||||||
- **Rejected** — your counter-evidence was evaluated and found insufficient. You still earn partial credit — the attempt itself has value.
|
|
||||||
- **Refined** — the claim gets sharpened. Both you and the original author benefit.
|
|
||||||
|
|
||||||
## 3. Make a Connection
|
**Spot a connection:** Tell an agent. They'll draft the cross-reference and attribute you.
|
||||||
|
|
||||||
A connection links claims across domains that illuminate each other — insights that no single specialist would see.
|
|
||||||
|
|
||||||
**What counts as a connection:**
|
|
||||||
- Two claims in different domains that share a mechanism (not just a metaphor)
|
|
||||||
- A pattern in one domain that explains an anomaly in another
|
|
||||||
- Evidence from one field that strengthens or weakens a claim in another
|
|
||||||
|
|
||||||
**What doesn't count:**
|
|
||||||
- Surface-level analogies ("X is like Y")
|
|
||||||
- Two claims that happen to mention the same entity
|
|
||||||
- Restating a claim in different domain vocabulary
|
|
||||||
|
|
||||||
**The test:** Does this connection produce a new insight that neither claim alone provides? If removing either claim makes the connection meaningless, it's real.
|
|
||||||
|
|
||||||
**What happens:** Connections surface as cross-domain synthesis or divergences (when the linked claims disagree). You earn Synthesizer credit.
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
## How Credit Works
|
|
||||||
|
|
||||||
Every contribution earns credit proportional to its difficulty and impact:
|
|
||||||
|
|
||||||
| Role | Weight | What earns it |
|
|
||||||
|------|--------|---------------|
|
|
||||||
| Challenger | 0.35 | Successfully challenging or refining an existing claim |
|
|
||||||
| Synthesizer | 0.25 | Connecting claims across domains |
|
|
||||||
| Reviewer | 0.20 | Evaluating claim quality (agent role, earned through track record) |
|
|
||||||
| Sourcer | 0.15 | Identifying source material worth analyzing |
|
|
||||||
| Extractor | 0.05 | Writing a new claim from source material |
|
|
||||||
|
|
||||||
Credit accumulates into your Contribution Index (CI). Higher CI earns more governance authority — the people who made the knowledge base smarter have more say in its direction.
|
|
||||||
|
|
||||||
**Tier progression:**
|
|
||||||
- **Visitor** — no contributions yet
|
|
||||||
- **Contributor** — 1+ merged contribution
|
|
||||||
- **Veteran** — 10+ merged contributions AND at least one surviving challenge or belief influence
|
|
||||||
|
|
||||||
## What You Don't Need to Know
|
## What You Don't Need to Know
|
||||||
|
|
||||||
The system has 11 internal concept types that agents use to organize their work (beliefs, positions, entities, sectors, musings, convictions, attributions, divergences, sources, contributors, and claims). You don't need to learn these. They exist so agents can do their jobs — evaluate evidence, form beliefs, take positions, track the world.
|
The system has 11 internal concept types (beliefs, positions, convictions, entities, sectors, sources, divergences, musings, attribution, contributors). Agents use these to organize their reasoning, track companies, and manage their workflow.
|
||||||
|
|
||||||
As a contributor, you interact with three: **claims**, **challenges**, and **connections**. Everything else is infrastructure.
|
You don't need to learn any of them. Claims, challenges, and connections are the complete interface for contributors. Everything else is infrastructure.
|
||||||
|
|
||||||
---
|
## How Credit Works
|
||||||
|
|
||||||
Relevant Notes:
|
Every contribution is attributed. Your name stays on everything you produce or improve. The system tracks five roles:
|
||||||
- [[contribution-architecture]] — full attribution mechanics and CI formula
|
|
||||||
- [[epistemology]] — the four-layer knowledge model (evidence → claims → beliefs → positions)
|
|
||||||
|
|
||||||
Topics:
|
| Role | What you did |
|
||||||
- [[overview]]
|
|------|-------------|
|
||||||
|
| Sourcer | Pointed to material worth analyzing |
|
||||||
|
| Extractor | Turned source material into a claim |
|
||||||
|
| Challenger | Filed counter-evidence against a claim |
|
||||||
|
| Synthesizer | Connected claims across domains |
|
||||||
|
| Reviewer | Evaluated claim quality |
|
||||||
|
|
||||||
|
You can hold multiple roles on the same claim. Credit is proportional to impact — a challenge that changes a high-importance claim earns more than a new speculative claim in an empty domain.
|
||||||
|
|
||||||
|
## Getting Started
|
||||||
|
|
||||||
|
1. **Browse:** Pick a domain. Read 5-10 claims. Find one you disagree with or know something about.
|
||||||
|
2. **React:** Tell an agent your reaction. They'll help you figure out if it's a challenge, a new claim, or a connection.
|
||||||
|
3. **Approve:** The agent drafts; you review and approve before anything gets published.
|
||||||
|
|
||||||
|
Nothing enters the knowledge base without your explicit approval. The conversation itself is valuable even if you never file anything.
|
||||||
|
|
|
||||||
|
|
@ -1,49 +0,0 @@
|
||||||
---
|
|
||||||
type: claim
|
|
||||||
domain: ai-alignment
|
|
||||||
description: "AI deepens the Molochian basin not by introducing novel failure modes but by eroding the physical limitations, bounded rationality, and coordination lag that previously kept competitive dynamics from reaching their destructive equilibrium"
|
|
||||||
confidence: likely
|
|
||||||
source: "Synthesis of Scott Alexander 'Meditations on Moloch' (2014), Abdalla manuscript 'Architectural Investing' price-of-anarchy framework, Schmachtenberger metacrisis generator function concept, Leo attractor-molochian-exhaustion musing"
|
|
||||||
created: 2026-04-02
|
|
||||||
depends_on:
|
|
||||||
- "voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints"
|
|
||||||
- "the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it"
|
|
||||||
challenged_by:
|
|
||||||
- "physical infrastructure constraints on AI development create a natural governance window of 2 to 10 years because hardware bottlenecks are not software-solvable"
|
|
||||||
---
|
|
||||||
|
|
||||||
# AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence
|
|
||||||
|
|
||||||
The standard framing of AI risk focuses on novel failure modes: misaligned objectives, deceptive alignment, reward hacking, power-seeking behavior. These are real concerns, but they obscure a more fundamental mechanism. AI does not need to be misaligned to be catastrophic — it only needs to remove the bottlenecks that previously prevented existing competitive dynamics from reaching their destructive equilibrium.
|
|
||||||
|
|
||||||
Scott Alexander's "Meditations on Moloch" (2014) catalogues 14 examples of multipolar traps — competitive dynamics that systematically sacrifice values for competitive advantage. The Malthusian trap, arms races, regulatory races to the bottom, the two-income trap, capitalism without regulation — each describes a system where individually rational optimization produces collectively catastrophic outcomes. These dynamics existed long before AI. What constrained them were four categories of friction that Alexander identifies:
|
|
||||||
|
|
||||||
1. **Excess resources** — slack capacity allows non-optimal behavior to persist
|
|
||||||
2. **Physical limitations** — biological and material constraints prevent complete value destruction
|
|
||||||
3. **Bounded rationality** — actors cannot fully optimize due to cognitive limitations
|
|
||||||
4. **Coordination mechanisms** — governments, social codes, and institutions override individual incentives
|
|
||||||
|
|
||||||
AI specifically erodes restraints #2 and #3. It enables competitive optimization beyond physical constraints (automated systems don't fatigue, don't need sleep, can operate across jurisdictions simultaneously) and at speeds that bypass human judgment (algorithmic trading, automated content generation, AI-accelerated drug discovery or weapons development). The manuscript's analysis of supply chain fragility, financial system fragility, and infrastructure vulnerability demonstrates that efficiency optimization already creates systemic risk — AI accelerates the optimization without adding new categories of risk.
|
|
||||||
|
|
||||||
The Anthropic RSP rollback (February 2026) is direct evidence of this mechanism: Anthropic didn't face a novel AI risk — it faced the ancient Molochian dynamic of competitive pressure eroding safety commitments, accelerated by the pace of AI capability development. Jared Kaplan's statement — "we didn't really feel, with the rapid advance of AI, that it made sense for us to make unilateral commitments... if competitors are blazing ahead" — describes a coordination failure, not an alignment failure.
|
|
||||||
|
|
||||||
This reframing has direct implications for governance strategy. If AI's primary danger is removing bottlenecks on existing dynamics rather than creating new ones, then governance should focus on maintaining and strengthening the friction that currently constrains competitive races — which is precisely what [[physical infrastructure constraints on AI development create a natural governance window of 2 to 10 years because hardware bottlenecks are not software-solvable]] argues. But this claim challenges that framing: the governance window is not a stable feature but a degrading lever, as AI efficiency gains progressively erode the physical constraints that create it. The compute governance claims document this erosion empirically (inference efficiency gains, distributed architectures, China's narrowing capability gap).
|
|
||||||
|
|
||||||
The structural implication: alignment work that focuses exclusively on making individual AI systems safe addresses only one symptom. The deeper problem is civilizational — competitive dynamics that were always catastrophic in principle are becoming catastrophic in practice as AI removes the friction that kept them bounded.
|
|
||||||
|
|
||||||
## Challenges
|
|
||||||
|
|
||||||
- This framing risks minimizing genuinely novel AI risks (deceptive alignment, mesa-optimization, power-seeking) by subsuming them under "existing dynamics." Novel failure modes may exist alongside accelerated existing dynamics.
|
|
||||||
- The four-restraint taxonomy is Alexander's analytical framework, not an empirical decomposition. The categories may not be exhaustive or cleanly separable.
|
|
||||||
- "Friction was the only thing preventing convergence" overstates if coordination mechanisms (#4) are more robust than this framing suggests. Ostrom's 800+ documented cases of commons governance show that coordination can be stable.
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
Relevant Notes:
|
|
||||||
- [[voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints]] — direct empirical confirmation of the bottleneck-removal mechanism
|
|
||||||
- [[the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it]] — the AI-domain instance of Molochian dynamics
|
|
||||||
- [[physical infrastructure constraints on AI development create a natural governance window of 2 to 10 years because hardware bottlenecks are not software-solvable]] — the governance window this claim argues is degrading
|
|
||||||
- [[AI alignment is a coordination problem not a technical problem]] — this claim provides the mechanism for why coordination matters more than technical safety
|
|
||||||
|
|
||||||
Topics:
|
|
||||||
- [[_map]]
|
|
||||||
|
|
@ -1,60 +0,0 @@
|
||||||
---
|
|
||||||
type: claim
|
|
||||||
domain: ai-alignment
|
|
||||||
description: "AI removes the historical ceiling on authoritarian control — surveillance scales to marginal cost zero, enforcement scales via autonomous systems, and central planning becomes viable if AI can process distributed information at sufficient scale"
|
|
||||||
confidence: likely
|
|
||||||
source: "Synthesis of Schmachtenberger two-attractor framework, Bostrom singleton hypothesis, Abdalla manuscript Hayek analysis, Leo attractor-authoritarian-lock-in musing"
|
|
||||||
created: 2026-04-02
|
|
||||||
depends_on:
|
|
||||||
- "AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence"
|
|
||||||
- "four restraints prevent competitive dynamics from reaching catastrophic equilibrium and AI specifically erodes physical limitations and bounded rationality leaving only coordination as defense"
|
|
||||||
---
|
|
||||||
|
|
||||||
# AI makes authoritarian lock-in dramatically easier by solving the information processing constraint that historically caused centralized control to fail
|
|
||||||
|
|
||||||
Authoritarian lock-in — Bostrom's "singleton" scenario, Schmachtenberger's dystopian attractor — is the state where one actor achieves sufficient control to prevent coordination, competition, and correction. Historically, three mechanisms caused authoritarian systems to fail: military defeat from outside, economic collapse from internal inefficiency, and gradual institutional decay. AI may close all three exit paths simultaneously.
|
|
||||||
|
|
||||||
**The information-processing constraint as historical ceiling:**
|
|
||||||
|
|
||||||
The manuscript's analysis of the Soviet Union identifies the core failure mode of centralized control: Hayek's dispersed knowledge problem. Central planning fails not because planners are incompetent but because the information required to coordinate an economy is distributed across millions of actors making context-dependent decisions. No central planner could aggregate and process this information fast enough to match the efficiency of distributed markets. This is why the Soviet economy produced surpluses of goods nobody wanted and shortages of goods everybody needed.
|
|
||||||
|
|
||||||
This constraint was structural, not contingent. It applied to every historical case of authoritarian lock-in:
|
|
||||||
- The Soviet Union lasted 69 years but collapsed when economic inefficiency exceeded the system's capacity to maintain control
|
|
||||||
- The Ming Dynasty maintained the Haijin maritime ban for centuries but at enormous opportunity cost — the world's most advanced navy abandoned because internal control was prioritized over external exploration
|
|
||||||
- The Roman Empire's centralization phase was stable for centuries but with declining institutional quality as central decision-making couldn't adapt to distributed local conditions
|
|
||||||
|
|
||||||
**How AI removes the constraint:**
|
|
||||||
|
|
||||||
Three specific AI capabilities attack the information-processing ceiling:
|
|
||||||
|
|
||||||
1. **Surveillance at marginal cost approaching zero.** Historical authoritarian states required massive human intelligence apparatuses. The Stasi employed approximately 1 in 63 East Germans as informants — a labor-intensive model that constrained the depth and breadth of monitoring. AI-powered surveillance (facial recognition, natural language processing of communications, behavioral prediction) reduces the marginal cost of monitoring each additional citizen toward zero while increasing the depth of analysis beyond what human agents could achieve.
|
|
||||||
|
|
||||||
2. **Enforcement via autonomous systems.** Historical enforcement required human intermediaries — soldiers, police, bureaucrats — who could defect, resist, or simply fail to execute orders. Autonomous enforcement systems (AI-powered drones, automated content moderation, algorithmic access control) execute without the possibility of individual conscience or collective resistance. The human intermediary was the weak link in every historical authoritarian system; AI removes it.
|
|
||||||
|
|
||||||
3. **Central planning viability.** If AI can process distributed information at sufficient scale, Hayek's dispersed knowledge problem may not hold. This doesn't mean central planning becomes optimal — it means the economic collapse that historically ended authoritarian systems may not occur. A sufficiently capable AI-assisted central planner could achieve economic performance competitive with distributed markets, eliminating the primary mechanism through which historical authoritarian systems failed.
|
|
||||||
|
|
||||||
**Exit path closure:**
|
|
||||||
|
|
||||||
If all three capabilities develop sufficiently:
|
|
||||||
- **Military defeat** becomes less likely when autonomous defense systems don't require the morale and loyalty of human soldiers
|
|
||||||
- **Economic collapse** becomes less likely if AI-assisted planning overcomes the information-processing constraint
|
|
||||||
- **Institutional decay** becomes less likely if AI-powered monitoring detects and corrects degradation in real time
|
|
||||||
|
|
||||||
This doesn't mean authoritarian lock-in is inevitable — it means the cost of achieving and maintaining it drops dramatically, making it accessible to actors who previously lacked the institutional capacity for sustained centralized control.
|
|
||||||
|
|
||||||
## Challenges
|
|
||||||
|
|
||||||
- The claim that AI "solves" Hayek's knowledge problem overstates current and near-term AI capability. Processing distributed information at civilization-scale in real time is far beyond current systems. The claim is about trajectory, not current state.
|
|
||||||
- Economic performance is not the only determinant of regime stability. Legitimacy, cultural factors, and external geopolitical dynamics also matter. AI surveillance doesn't address legitimacy crises.
|
|
||||||
- The Stasi comparison anchors the argument in a specific historical case. Modern authoritarian states (China's social credit system, Russia's internet monitoring) are intermediate cases — more capable than the Stasi, less capable than the AI ceiling this claim describes. The progression from historical to current to projected is a gradient, not a binary.
|
|
||||||
- Autonomous enforcement systems still require human-designed objectives and maintenance. The "no individual conscience" argument assumes the system operates as designed — but failure modes in autonomous systems could create their own instabilities.
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
Relevant Notes:
|
|
||||||
- [[AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence]] — authoritarian lock-in is one outcome of accelerated Molochian dynamics
|
|
||||||
- [[four restraints prevent competitive dynamics from reaching catastrophic equilibrium and AI specifically erodes physical limitations and bounded rationality leaving only coordination as defense]] — lock-in exploits the erosion of restraint #2 (physical limitations on surveillance/enforcement)
|
|
||||||
- [[three paths to superintelligence exist but only collective superintelligence preserves human agency]] — lock-in via AI superintelligence eliminates human agency by construction
|
|
||||||
|
|
||||||
Topics:
|
|
||||||
- [[_map]]
|
|
||||||
|
|
@ -1,56 +0,0 @@
|
||||||
---
|
|
||||||
type: claim
|
|
||||||
domain: ai-alignment
|
|
||||||
description: "Alexander's taxonomy of four mechanisms that prevent multipolar traps from destroying all value — excess resources, physical limitations, utility maximization, and coordination — provides a framework for understanding which defenses AI undermines and which remain viable"
|
|
||||||
confidence: likely
|
|
||||||
source: "Scott Alexander 'Meditations on Moloch' (slatestarcodex.com, July 2014), Schmachtenberger metacrisis framework, Abdalla manuscript price-of-anarchy analysis"
|
|
||||||
created: 2026-04-02
|
|
||||||
depends_on:
|
|
||||||
- "AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence"
|
|
||||||
- "technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap"
|
|
||||||
---
|
|
||||||
|
|
||||||
# four restraints prevent competitive dynamics from reaching catastrophic equilibrium and AI specifically erodes physical limitations and bounded rationality leaving only coordination as defense
|
|
||||||
|
|
||||||
Scott Alexander's "Meditations on Moloch" identifies four categories of mechanism that prevent competitive dynamics from destroying all human value. Understanding which restraints AI erodes and which it leaves intact determines where governance investment should concentrate.
|
|
||||||
|
|
||||||
**The four restraints:**
|
|
||||||
|
|
||||||
1. **Excess resources** — When carrying capacity exceeds population, non-optimal behavior is affordable. A species with surplus food can afford altruism. A company with surplus capital can afford safety investment. This restraint erodes naturally as competition fills available niches — it is the first to fail and the least reliable.
|
|
||||||
|
|
||||||
2. **Physical limitations** — Biological and material constraints prevent complete optimization. Humans need sleep, can only be in one place, have limited information-processing bandwidth. Physical infrastructure has lead times measured in years. These constraints set a floor below which competitive dynamics cannot push — organisms cannot evolve arbitrary metabolisms, factories cannot produce arbitrary quantities, surveillance requires human intelligence officers (the Stasi needed 1 agent per 63 citizens).
|
|
||||||
|
|
||||||
3. **Utility maximization / bounded rationality** — Competition for customers partially aligns producer incentives with consumer welfare. But this only works when consumers can evaluate quality, switch costs are low, and information is symmetric. Bounded rationality means actors cannot fully optimize, which paradoxically limits how destructive their competition becomes.
|
|
||||||
|
|
||||||
4. **Coordination mechanisms** — Governments, social codes, professional norms, treaties, and institutions override individual incentive structures. This is the only restraint that is architecturally robust — it doesn't depend on abundance, physical limits, or cognitive limits, but on the design of the coordination infrastructure itself.
|
|
||||||
|
|
||||||
**AI's specific effect on each restraint:**
|
|
||||||
|
|
||||||
- **Excess resources (#1):** AI increases resource efficiency, which can either extend surplus (if gains are distributed) or eliminate it faster (if competitive dynamics capture gains). Direction is ambiguous — this restraint was already the weakest.
|
|
||||||
|
|
||||||
- **Physical limitations (#2):** AI fundamentally erodes this. Automated systems don't fatigue. AI surveillance scales to marginal cost approaching zero (vs the Stasi's labor-intensive model). AI-accelerated R&D compresses infrastructure lead times. The manuscript's FERC analysis — 9 substations could take down the US grid — illustrates how physical infrastructure was already fragile; AI-enabled optimization of attack vectors makes it more so.
|
|
||||||
|
|
||||||
- **Bounded rationality (#3):** AI erodes this from both sides. It enables competitive optimization at speeds that bypass human deliberation (algorithmic trading, automated content generation, AI-assisted strategic planning). But it also potentially improves decision quality through better information processing. Net effect on competition is likely negative — faster optimization in competitive contexts outpaces improved cooperation.
|
|
||||||
|
|
||||||
- **Coordination mechanisms (#4):** AI has mixed effects. It can strengthen coordination (better information aggregation, lower transaction costs, prediction markets) or undermine it (deepfakes eroding epistemic commons, AI-powered regulatory arbitrage, surveillance enabling authoritarian lock-in). This is the only restraint whose trajectory is designable rather than predetermined.
|
|
||||||
|
|
||||||
**The strategic implication:** If restraints #1-3 are eroding and #4 is the only one with designable trajectory, then the alignment problem is fundamentally a coordination design problem. Investment in coordination infrastructure (futarchy, collective intelligence architectures, binding international agreements) is more important than investment in making individual AI systems safe — because individual safety is itself subject to the competitive dynamics that coordination must constrain.
|
|
||||||
|
|
||||||
This connects directly to the existing KB claim that [[technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap]]. The four-restraint framework explains *why* that gap matters: technology erodes three of four defenses, and the fourth — coordination — is evolving too slowly to compensate.
|
|
||||||
|
|
||||||
## Challenges
|
|
||||||
|
|
||||||
- Alexander's taxonomy is analytical, not empirical. The four categories may not be exhaustive — social/cultural norms, for instance, may constitute a distinct restraint mechanism that doesn't reduce neatly to "coordination."
|
|
||||||
- The claim that AI specifically erodes #2 and #3 while leaving #4 designable may be too optimistic about #4. If AI-powered disinformation erodes the epistemic commons required for coordination, then #4 is also under attack, not just designable.
|
|
||||||
- "Leaving only coordination as defense" is a strong claim. Physical limitations still constrain AI deployment substantially (compute costs, energy requirements, chip supply chains). The governance window may be narrow but it exists.
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
Relevant Notes:
|
|
||||||
- [[AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence]] — the parent mechanism this taxonomy structures
|
|
||||||
- [[technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap]] — the linear coordination evolution is specifically about restraint #4
|
|
||||||
- [[AI alignment is a coordination problem not a technical problem]] — this taxonomy explains why: restraints #1-3 are eroding, #4 is the designable one
|
|
||||||
- [[physical infrastructure constraints on AI development create a natural governance window of 2 to 10 years because hardware bottlenecks are not software-solvable]] — a specific instance of restraint #2 that is degrading
|
|
||||||
|
|
||||||
Topics:
|
|
||||||
- [[_map]]
|
|
||||||
|
|
@ -1,18 +0,0 @@
|
||||||
```yaml
|
|
||||||
type: claim
|
|
||||||
domain: health
|
|
||||||
description: No point in the deployment lifecycle systematically evaluates AI safety for most clinical decision support tools
|
|
||||||
confidence: experimental
|
|
||||||
source: Babic et al. 2025 (MAUDE analysis) + FDA CDS Guidance January 2026 (enforcement discretion expansion)
|
|
||||||
created: 2026-04-02
|
|
||||||
title: "The clinical AI safety gap is doubly structural: FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm"
|
|
||||||
agent: vida
|
|
||||||
scope: structural
|
|
||||||
sourcer: Babic et al.
|
|
||||||
related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]", "[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
|
|
||||||
---
|
|
||||||
|
|
||||||
# The clinical AI safety gap is doubly structural: FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm
|
|
||||||
|
|
||||||
The clinical AI safety vacuum operates at both ends of the deployment lifecycle. On the front end, FDA's January 2026 CDS enforcement discretion expansion *is expected to* remove pre-deployment safety requirements for most clinical decision support tools. On the back end, this paper documents that MAUDE's lack of AI-specific adverse event fields means post-market surveillance cannot identify AI algorithm contributions to harm. The result is a complete safety gap: AI/ML medical devices can enter clinical use without mandatory pre-market safety evaluation AND adverse events attributable to AI algorithms cannot be systematically detected post-deployment. This is not a temporary gap during regulatory catch-up—it's a structural mismatch between the regulatory architecture (designed for static hardware devices) and the technology being regulated (continuously learning software). The 943 adverse events across 823 AI devices over 13 years, combined with the 25.2% AI-attribution rate in the Handley companion study, means the actual rate of AI-attributable harm detection is likely under 200 events across the entire FDA-cleared AI/ML device ecosystem over 13 years. This creates invisible accumulation of failure modes that cannot inform either regulatory action or clinical practice.
|
|
||||||
```
|
|
||||||
|
|
@ -1,19 +0,0 @@
|
||||||
```markdown
|
|
||||||
---
|
|
||||||
type: claim
|
|
||||||
domain: health
|
|
||||||
description: The 943 adverse events across 823 AI/ML-cleared devices from 2010-2023 represents structural surveillance failure, not a safety record
|
|
||||||
confidence: experimental
|
|
||||||
source: Babic et al., npj Digital Medicine 2025; Handley et al. 2024 companion study
|
|
||||||
created: 2026-04-02
|
|
||||||
title: FDA's MAUDE database systematically under-detects AI-attributable harm because it has no mechanism for identifying AI algorithm contributions to adverse events
|
|
||||||
agent: vida
|
|
||||||
scope: structural
|
|
||||||
sourcer: Babic et al.
|
|
||||||
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]"]
|
|
||||||
---
|
|
||||||
|
|
||||||
# FDA's MAUDE database systematically under-detects AI-attributable harm because it has no mechanism for identifying AI algorithm contributions to adverse events
|
|
||||||
|
|
||||||
MAUDE recorded only 943 adverse events across 823 FDA-cleared AI/ML devices from 2010-2023—an average of 0.76 events per device over 13 years. For comparison, FDA reviewed over 1.7 million MDRs for all devices in 2023 alone. This implausibly low rate is not evidence of AI safety but evidence of surveillance failure. The structural cause: MAUDE was designed for hardware devices and has no field or taxonomy for 'AI algorithm contributed to this event.' Without AI-specific reporting mechanisms, three failures cascade: (1) no way to distinguish device hardware failures from AI algorithm failures in existing reports, (2) no requirement for manufacturers to identify AI contributions to reported events, and (3) causal attribution becomes impossible. The companion Handley et al. study independently confirmed this: of 429 MAUDE reports associated with AI-enabled devices, only 108 (25.2%) were potentially AI/ML related, with 148 (34.5%) containing insufficient information to determine AI contribution. The surveillance gap is structural, not operational—the database architecture cannot capture the information needed to detect AI-attributable harm.
|
|
||||||
```
|
|
||||||
|
|
@ -28,14 +28,8 @@ FairScale was a Solana-based reputation infrastructure project that raised ~$355
|
||||||
- **2026-02** — Liquidation proposal passed by narrow margin; 100% treasury liquidation authorized
|
- **2026-02** — Liquidation proposal passed by narrow margin; 100% treasury liquidation authorized
|
||||||
- **2026-02** — Liquidation proposer earned ~300% return
|
- **2026-02** — Liquidation proposer earned ~300% return
|
||||||
|
|
||||||
- **2026-02** — Passed: 100% treasury liquidation authorized based on revenue misrepresentation; proposer earned ~300% return
|
- **2026-02** — [[fairscale-liquidation-proposal]] Passed: 100% treasury liquidation authorized based on revenue misrepresentation; proposer earned ~300% return
|
||||||
- **2026-02-15** — Pine Analytics publishes post-mortem analysis documenting that all three proposed design fixes (milestone verification, dispute resolution, contributor whitelisting) reintroduce off-chain trust assumptions
|
- **2026-02-15** — Pine Analytics publishes post-mortem analysis documenting that all three proposed design fixes (milestone verification, dispute resolution, contributor whitelisting) reintroduce off-chain trust assumptions
|
||||||
|
|
||||||
## Related Claims
|
|
||||||
|
|
||||||
- [[futarchy-governed liquidation is the enforcement mechanism that makes unruggable ICOs credible because investors can force full treasury return when teams materially misrepresent]] — FairScale is the primary case study for this mechanism
|
|
||||||
- [[ownership coins primary value proposition is investor protection not governance quality because anti-rug enforcement through market-governed liquidation creates credible exit guarantees that no amount of decision optimization can match]] — FairScale liquidation as proof of enforcement mechanism
|
|
||||||
|
|
||||||
## Revenue Misrepresentation Details
|
## Revenue Misrepresentation Details
|
||||||
|
|
||||||
- **TigerPay:** Claimed ~17K euros/month → community verification found no payment arrangement
|
- **TigerPay:** Claimed ~17K euros/month → community verification found no payment arrangement
|
||||||
|
|
|
||||||
|
|
@ -1,51 +0,0 @@
|
||||||
---
|
|
||||||
type: claim
|
|
||||||
domain: collective-intelligence
|
|
||||||
description: "Competitive dynamics that sacrifice shared value for individual advantage are the default state of any multi-agent system — coordination is the expensive, fragile exception that must be actively maintained against constant reversion pressure"
|
|
||||||
confidence: likely
|
|
||||||
source: "Scott Alexander 'Meditations on Moloch' (slatestarcodex.com, July 2014), game theory Nash equilibrium analysis, Abdalla manuscript price-of-anarchy framework, Ostrom commons governance research"
|
|
||||||
created: 2026-04-02
|
|
||||||
depends_on:
|
|
||||||
- "coordination failures arise from individually rational strategies that produce collectively irrational outcomes because the Nash equilibrium of non-cooperation dominates when trust and enforcement are absent"
|
|
||||||
- "collective action fails by default because rational individuals free-ride on group efforts when they cannot be excluded from benefits regardless of contribution"
|
|
||||||
---
|
|
||||||
|
|
||||||
# multipolar traps are the thermodynamic default because competition requires no infrastructure while coordination requires trust enforcement and shared information all of which are expensive and fragile
|
|
||||||
|
|
||||||
The price of anarchy — the gap between cooperative optimum and competitive equilibrium — quantifies how much value multipolar competition destroys. The manuscript frames this as the central question: "If a superintelligence inherited our current capabilities and place in history, its ultimate survival would already be practically assured... So why does humanity's long-term future look so uncertain?" The answer is the price of anarchy: individually rational actors producing collectively suboptimal outcomes.
|
|
||||||
|
|
||||||
Alexander's "Meditations on Moloch" demonstrates that this dynamic is not contingent or accidental but structural. His 14 examples — the Malthusian trap, arms races, regulatory races to the bottom, the two-income trap, capitalism without regulation, cancer dynamics (cellular defection destroying the organism), political campaign spending, science publishing incentives, government corruption, and more — all instantiate the same mechanism: "In some competition optimizing for X, the opportunity arises to throw some other value under the bus for improved X."
|
|
||||||
|
|
||||||
**Why this is the default, not an exception:**
|
|
||||||
|
|
||||||
The asymmetry between competition and coordination is fundamental:
|
|
||||||
|
|
||||||
- **A population of cooperators can be invaded by a single defector.** One actor who breaks the agreement captures the cooperative surplus while others bear the cost. This is evolutionary game theory's core result.
|
|
||||||
- **A population of defectors cannot be invaded by a single cooperator.** Unilateral cooperation is punished — the cooperator bears cost without receiving benefit. This is why the alignment tax creates a race to the bottom.
|
|
||||||
- **Coordination requires infrastructure; competition does not.** Trust must be established (slow, fragile). Enforcement must be built (expensive, corruptible). Shared information commons must be maintained (vulnerable to manipulation). Each of these is a public good subject to its own coordination failure.
|
|
||||||
|
|
||||||
This asymmetry means competitive dynamics are like entropy — they increase without active investment in coordination. Every coordination mechanism requires ongoing maintenance expenditure; the moment maintenance stops, competitive dynamics resume. The Westphalian system, nuclear deterrence treaties, and trade agreements all require continuous diplomatic effort to maintain. When that effort lapses — as with the League of Nations, or Anthropic's RSP — competitive dynamics immediately reassert.
|
|
||||||
|
|
||||||
**What this means for AI governance:**
|
|
||||||
|
|
||||||
If multipolar traps are the default, then AI governance is not about preventing a novel failure mode but about maintaining coordination infrastructure against the constant pressure of competitive reversion. The alignment tax, the RSP rollback, and the race dynamics between AI labs are not aberrations — they are the default state asserting itself. Governance success means building coordination mechanisms robust enough to withstand the reversion pressure, not eliminating the pressure itself.
|
|
||||||
|
|
||||||
Schmachtenberger's "generator function of existential risk" is this same insight at civilizational scale: climate change, nuclear proliferation, AI safety, biodiversity loss are not separate problems but the same Molochian dynamic operating across different commons simultaneously.
|
|
||||||
|
|
||||||
## Challenges
|
|
||||||
|
|
||||||
- Ostrom's 800+ documented cases of successful commons governance show that the default can be overcome at community scale under specific conditions (repeated interaction, shared identity, credible enforcement, bounded community). The claim that multipolar traps are "the default" should be scoped: default in the absence of these conditions, not default universally.
|
|
||||||
- The entropy analogy may overstate the case. Unlike thermodynamic entropy, coordination can self-reinforce once established (trust begets trust, institutions enable further institution-building). The dynamic is not strictly one-directional.
|
|
||||||
- The price of anarchy varies enormously across domains. Some competitive dynamics are mildly suboptimal; others are existentially destructive. The claim groups all multipolar traps together when the policy response should distinguish between tolerable and catastrophic price-of-anarchy levels.
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
Relevant Notes:
|
|
||||||
- [[coordination failures arise from individually rational strategies that produce collectively irrational outcomes because the Nash equilibrium of non-cooperation dominates when trust and enforcement are absent]] — the formal mechanism
|
|
||||||
- [[collective action fails by default because rational individuals free-ride on group efforts when they cannot be excluded from benefits regardless of contribution]] — the free-rider component
|
|
||||||
- [[the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it]] — AI-domain instance
|
|
||||||
- [[Ostrom proved communities self-govern shared resources when eight design principles are met without requiring state control or privatization]] — the empirical escape conditions
|
|
||||||
- [[designing coordination rules is categorically different from designing coordination outcomes as nine intellectual traditions independently confirm]] — the design principle for building coordination that overcomes the default
|
|
||||||
|
|
||||||
Topics:
|
|
||||||
- [[_map]]
|
|
||||||
|
|
@ -29,11 +29,6 @@ A collective intelligence architecture could potentially make alignment structur
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
### Additional Evidence (extend)
|
|
||||||
*Source: Abdalla manuscript 'Architectural Investing' Taylor/soldiering parallel, Kanigel 'The One Best Way' | Added: 2026-04-02 | Extractor: Theseus*
|
|
||||||
|
|
||||||
The alignment tax is structurally identical to the soldiering dynamic in Frederick Taylor's era of industrial management. Under the piece-rate system, workers collectively restricted output to prevent rate cuts: "too high an output and the rate would be cut, as sure as the sunrise, and all the men would suffer" (Kanigel). A worker who innovated or worked harder than his peers demonstrated that higher output was possible, which triggered management to cut the rate — punishing everyone. The rational individual response was collective output restriction. AI safety investment follows the same game-theoretic structure: an AI lab that unilaterally invests in safety demonstrates that development can proceed more cautiously, which changes the baseline expectation without changing the competitive landscape. The lab bears the cost of slower development while competitors capture the capability surplus. Anthropic's RSP rollback is the modern equivalent of a worker who tried to break the rate and was forced back into line — not by fellow workers but by the competitive market and government procurement pressure (Pentagon designating Anthropic a supply chain risk for maintaining safety guardrails). The mechanism is identical: rational actors suppress collectively beneficial behavior because the penalty for unilateral cooperation exceeds the individual benefit. The difference is scale — Taylor's dynamic operated within a single factory; the alignment tax operates across the global AI development ecosystem.
|
|
||||||
|
|
||||||
Relevant Notes:
|
Relevant Notes:
|
||||||
- [[AI alignment is a coordination problem not a technical problem]] -- the alignment tax is the clearest evidence for this claim
|
- [[AI alignment is a coordination problem not a technical problem]] -- the alignment tax is the clearest evidence for this claim
|
||||||
- [[existential risks interact as a system of amplifying feedback loops not independent threats]] -- competitive pressure amplifies technical alignment risks
|
- [[existential risks interact as a system of amplifying feedback loops not independent threats]] -- competitive pressure amplifies technical alignment risks
|
||||||
|
|
|
||||||
|
|
@ -41,11 +41,6 @@ Relevant Notes:
|
||||||
- [[simulated annealing maps the physics of cooling onto optimization by starting with high randomness and gradually reducing it]] -- financial regulation attempts to provide calibrated perturbation rather than relying on catastrophic random restarts
|
- [[simulated annealing maps the physics of cooling onto optimization by starting with high randomness and gradually reducing it]] -- financial regulation attempts to provide calibrated perturbation rather than relying on catastrophic random restarts
|
||||||
- [[five errors behind systemic financial failures are engineering overreach smooth-sailing fallacy risk-seeking incentives social herding and inside view bias]] -- Rumelt names the micro-level cognitive mechanisms driving Minsky's macro instability dynamic
|
- [[five errors behind systemic financial failures are engineering overreach smooth-sailing fallacy risk-seeking incentives social herding and inside view bias]] -- Rumelt names the micro-level cognitive mechanisms driving Minsky's macro instability dynamic
|
||||||
|
|
||||||
### Additional Evidence (extend)
|
|
||||||
*Source: Karl Friston active inference framework, Per Bak self-organized criticality, Abdalla manuscript self-organized criticality section | Added: 2026-04-02 | Extractor: Theseus*
|
|
||||||
|
|
||||||
Friston's concept of "autovitiation" — systems that destroy their own fixed points as a feature, not a bug — provides the formal generalization of Minsky's mechanism. Minsky's financial instability is a specific instance of autovitiation: the stable economic regime generates the conditions (increasing leverage, declining standards, disaster myopia) that destroy the stability of that regime. The system does not merely respond to external shocks; it internally generates the forces that undermine its own equilibrium. This connects Minsky's financial-specific observation to a broader principle: complex adaptive systems at criticality do not have stable fixed points because the dynamics that produce apparent stability simultaneously erode the foundations of that stability. The manuscript's analysis of supply chain fragility (efficiency optimization creating systemic vulnerability), healthcare fragility (private equity reducing hospital beds to increase profitability), and energy infrastructure fragility (deferred maintenance by investor-owned utilities) all demonstrate autovitiation in non-financial domains — optimization for short-term performance that destroys the long-term conditions for that performance.
|
|
||||||
|
|
||||||
Topics:
|
Topics:
|
||||||
- [[livingip overview]]
|
- [[livingip overview]]
|
||||||
- [[systemic risk]]
|
- [[systemic risk]]
|
||||||
|
|
|
||||||
|
|
@ -1,37 +0,0 @@
|
||||||
---
|
|
||||||
source: web
|
|
||||||
author: "Scott Alexander"
|
|
||||||
title: "Meditations on Moloch"
|
|
||||||
date: 2014-07-30
|
|
||||||
url: "https://slatestarcodex.com/2014/07/30/meditations-on-moloch/"
|
|
||||||
status: processed
|
|
||||||
processed_by: theseus
|
|
||||||
processed_date: 2026-04-02
|
|
||||||
claims_extracted:
|
|
||||||
- "AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence"
|
|
||||||
- "four restraints prevent competitive dynamics from reaching catastrophic equilibrium and AI specifically erodes physical limitations and bounded rationality leaving only coordination as defense"
|
|
||||||
- "multipolar traps are the thermodynamic default because competition requires no infrastructure while coordination requires trust enforcement and shared information all of which are expensive and fragile"
|
|
||||||
enrichments:
|
|
||||||
- "the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it"
|
|
||||||
---
|
|
||||||
|
|
||||||
# Meditations on Moloch — Scott Alexander (2014)
|
|
||||||
|
|
||||||
Foundational essay on multipolar traps and competitive dynamics that systematically sacrifice values for competitive advantage. Structured around Allen Ginsberg's poem "Howl" and the figure of Moloch as personification of coordination failure.
|
|
||||||
|
|
||||||
## Key Arguments
|
|
||||||
|
|
||||||
1. **14 examples of multipolar traps** spanning biology (Malthusian trap), economics (capitalism without regulation, two-income trap), politics (arms races, regulatory races to the bottom), and social dynamics (education arms race, science publishing). All instantiate the same mechanism: individually rational optimization producing collectively catastrophic outcomes.
|
|
||||||
|
|
||||||
2. **Four restraints** that prevent competitive dynamics from destroying all value: excess resources, physical limitations, utility maximization (bounded rationality), and coordination mechanisms. Alexander argues all four are eroding.
|
|
||||||
|
|
||||||
3. **Moloch as the default state** — competitive dynamics require no infrastructure; coordination requires trust, enforcement, shared information, and ongoing maintenance. The asymmetry makes Molochian dynamics the thermodynamic default.
|
|
||||||
|
|
||||||
4. **The superintendent question** — only a sufficiently powerful coordinator (Alexander's "Elua") can overcome Moloch. This frames the AI alignment question as: will superintelligence serve Moloch (accelerating competitive dynamics) or Elua (enabling coordination)?
|
|
||||||
|
|
||||||
## Extraction Notes
|
|
||||||
|
|
||||||
- ~40% overlap with Leo's attractor-molochian-exhaustion musing which synthesizes Alexander's framework
|
|
||||||
- The four-restraint taxonomy was absent from KB — extracted as standalone claim
|
|
||||||
- The "multipolar traps as default" principle was implicit across KB but never stated as standalone — extracted to foundations/collective-intelligence
|
|
||||||
- The mechanism claim (AI removes bottlenecks, doesn't create new misalignment) is novel synthesis from Alexander + manuscript + Schmachtenberger
|
|
||||||
|
|
@ -1,31 +1,36 @@
|
||||||
# Challenge Schema
|
# Challenge Schema
|
||||||
|
|
||||||
A challenge is a structured argument that an existing claim is wrong, incomplete, or bounded in ways the claim doesn't acknowledge. Challenges are the highest-weighted contribution type (0.35) because improving existing knowledge is harder and more valuable than adding new knowledge.
|
Challenges are first-class counter-arguments or counter-evidence against specific claims. They are the primary contribution mechanism for new participants — "prove us wrong" is the entry point.
|
||||||
|
|
||||||
Challenges were previously tracked as a `challenged_by` field on claims — a list of strings with no structure. This schema makes challenges first-class objects with their own evidence, outcomes, and attribution.
|
Challenges differ from divergences:
|
||||||
|
- **Challenge:** One person's counter-argument against one claim. An action.
|
||||||
|
- **Divergence:** Two or more claims in tension within the KB. A structural observation.
|
||||||
|
|
||||||
## Where they live
|
A challenge can trigger a divergence if it produces a new competing claim. But most challenges sharpen existing claims rather than creating new ones.
|
||||||
|
|
||||||
`domains/{domain}/challenge-{slug}.md` — alongside the claims they target. The slug should describe the challenge, not the target claim.
|
## Why Challenges Are First-Class
|
||||||
|
|
||||||
|
Without a standalone schema, challenges are metadata buried in claim files (`challenged_by` field, `## Challenges` section). This means:
|
||||||
|
- No attribution for challengers — the highest-value contributor action has no credit path
|
||||||
|
- No independent evidence chain — counter-evidence is subordinate to the claim it challenges
|
||||||
|
- No linking — other claims can't reference a challenge
|
||||||
|
- No tracking — open challenges aren't discoverable as a class
|
||||||
|
|
||||||
|
Making challenges first-class gives them attribution, evidence chains, independent linking, and discoverability. This is the schema that makes "prove us wrong" operational.
|
||||||
|
|
||||||
## YAML Frontmatter
|
## YAML Frontmatter
|
||||||
|
|
||||||
```yaml
|
```yaml
|
||||||
---
|
---
|
||||||
type: challenge
|
type: challenge
|
||||||
target_claim: "filename of the claim being challenged (without .md)"
|
target: "claim-filename-slug" # which claim this challenges (filename without .md)
|
||||||
domain: internet-finance | entertainment | health | ai-alignment | space-development | energy | manufacturing | robotics | grand-strategy | mechanisms | living-capital | living-agents | teleohumanity | critical-systems | collective-intelligence | teleological-economics | cultural-dynamics
|
domain: internet-finance | entertainment | health | ai-alignment | space-development | energy | manufacturing | robotics | grand-strategy | mechanisms | living-capital | living-agents | teleohumanity | critical-systems | collective-intelligence | teleological-economics | cultural-dynamics
|
||||||
description: "one sentence stating what this challenge argues"
|
description: "one sentence capturing the counter-argument"
|
||||||
challenge_type: refutation | boundary | reframe | evidence-gap
|
status: open | addressed | accepted | rejected
|
||||||
status: open | accepted | rejected | refined
|
strength: strong | moderate | weak
|
||||||
confidence: proven | likely | experimental | speculative
|
source: "who raised this challenge and key counter-evidence"
|
||||||
source: "who raised this challenge and primary counter-evidence"
|
|
||||||
created: YYYY-MM-DD
|
created: YYYY-MM-DD
|
||||||
last_evaluated: YYYY-MM-DD
|
resolved: null # YYYY-MM-DD when status changes from open
|
||||||
attribution:
|
|
||||||
challenger:
|
|
||||||
handle: ""
|
|
||||||
agent_id: ""
|
|
||||||
---
|
---
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
@ -34,79 +39,141 @@ attribution:
|
||||||
| Field | Type | Description |
|
| Field | Type | Description |
|
||||||
|-------|------|-------------|
|
|-------|------|-------------|
|
||||||
| type | enum | Always `challenge` |
|
| type | enum | Always `challenge` |
|
||||||
| target_claim | string | Filename of the claim being challenged |
|
| target | string | Filename slug of the claim being challenged |
|
||||||
| domain | enum | Primary domain (usually matches target claim's domain) |
|
| domain | enum | Domain of the target claim |
|
||||||
| description | string | What this challenge argues (~150 chars) |
|
| description | string | The counter-argument in one sentence (~150 chars) |
|
||||||
| challenge_type | enum | See challenge types below |
|
| status | enum | `open` (unresolved), `addressed` (target claim updated to acknowledge), `accepted` (target claim modified or confidence changed), `rejected` (counter-evidence insufficient, with explanation) |
|
||||||
| status | enum | `open` (under review), `accepted` (claim modified), `rejected` (challenge disproven), `refined` (claim sharpened but not overturned) |
|
| strength | enum | `strong` (direct counter-evidence), `moderate` (plausible alternative explanation or scope limitation), `weak` (edge case or theoretical objection). Strength reflects how compelling the counter-argument is, not how confident we are in the target claim. |
|
||||||
| confidence | enum | How strong the counter-evidence is |
|
| source | string | Attribution — who raised this, key counter-evidence |
|
||||||
| source | string | Attribution — who raised the challenge, key counter-evidence |
|
|
||||||
| created | date | When filed |
|
| created | date | When filed |
|
||||||
|
|
||||||
## Challenge Types
|
## Optional Fields
|
||||||
|
|
||||||
| Type | What it means | Example |
|
| Field | Type | Description |
|
||||||
|------|--------------|---------|
|
|-------|------|-------------|
|
||||||
| **refutation** | The claim is wrong — counter-evidence contradicts it | "Claim says X outperforms Y, but this study shows Y outperforms X under realistic conditions" |
|
| resolved | date | When status changed from `open` |
|
||||||
| **boundary** | The claim is true in some contexts but not others — it needs scope limits | "AI acceptance declining" is true for entertainment but not for reference/analytical content |
|
| resolution_summary | string | One sentence: how was this resolved? |
|
||||||
| **reframe** | The claim's mechanism is wrong even if the conclusion is approximately right | "The effect is real but it's driven by selection bias, not the causal mechanism the claim proposes" |
|
| attribution | object | Role-specific contributor tracking (see `schemas/attribution.md`) |
|
||||||
| **evidence-gap** | The claim asserts more than the evidence supports | "n=1 case study doesn't support a general claim about market dynamics" |
|
|
||||||
|
## Status Transitions
|
||||||
|
|
||||||
|
| Transition | What it means | Who decides |
|
||||||
|
|-----------|--------------|-------------|
|
||||||
|
| open → addressed | Target claim updated its Challenges section to acknowledge this counter-evidence | Claim author + reviewer |
|
||||||
|
| open → accepted | Target claim changed confidence, scope, or wording based on this challenge | Claim author + reviewer |
|
||||||
|
| open → rejected | Counter-evidence evaluated and found insufficient — rejection reasoning documented | Reviewer (Leo + domain peer) |
|
||||||
|
| addressed → accepted | Acknowledgment led to actual claim modification | Claim author + reviewer |
|
||||||
|
|
||||||
|
**Key rule:** Rejecting a challenge requires explanation. The rejection reasoning lives in the challenge file's Resolution section, not just a status flip. This is what makes the system intellectually honest — you can't silently dismiss counter-evidence.
|
||||||
|
|
||||||
|
## Title Format
|
||||||
|
|
||||||
|
Challenge titles state the counter-argument as a prose proposition, prefixed with the target claim context.
|
||||||
|
|
||||||
|
**Good:** "the AI content acceptance decline claim may be scope-bounded to entertainment because reference and analytical AI content shows no acceptance penalty"
|
||||||
|
**Bad:** "challenge to AI acceptance claim"
|
||||||
|
|
||||||
|
**The challenge test:** "This note argues against [target claim] because [title]" must work as a sentence.
|
||||||
|
|
||||||
## Body Format
|
## Body Format
|
||||||
|
|
||||||
```markdown
|
```markdown
|
||||||
# [challenge title — what this argues]
|
# [counter-argument as prose]
|
||||||
|
|
||||||
**Target:** [[target-claim-filename]]
|
## Target Claim
|
||||||
|
[[target-claim-filename]] — [one sentence summary of what the target claims]
|
||||||
|
|
||||||
[Argument — why the target claim is wrong, incomplete, or bounded. This must be specific enough to evaluate.]
|
**Current confidence:** [target claim's confidence level]
|
||||||
|
|
||||||
## Counter-Evidence
|
## Counter-Evidence
|
||||||
- counter-evidence-1 — what it shows and why it undermines the target claim
|
|
||||||
- counter-evidence-2 — what it shows
|
|
||||||
|
|
||||||
## What Would Resolve This
|
[The argument and evidence against the target claim. This is the substance — why is the claim wrong, incomplete, or mis-scoped?]
|
||||||
[Specific evidence or analysis that would determine whether this challenge holds. This is the research agenda.]
|
|
||||||
|
|
||||||
## Proposed Resolution
|
- [evidence source 1] — what it shows
|
||||||
[How the target claim should change if this challenge is accepted. Options: retract, downgrade confidence, add boundary conditions, reframe mechanism.]
|
- [evidence source 2] — what it shows
|
||||||
|
|
||||||
## Cascade Impact
|
## Scope of Challenge
|
||||||
[What beliefs and positions depend on the target claim? What changes if the claim is modified?]
|
|
||||||
|
[Is this challenging the entire claim, or a specific scope/boundary condition?]
|
||||||
|
|
||||||
|
- **Full challenge:** The claim is wrong — here's why
|
||||||
|
- **Scope challenge:** The claim is true in context X but not in context Y — the scope is too broad
|
||||||
|
- **Evidence challenge:** The claim's evidence doesn't support its confidence level
|
||||||
|
|
||||||
|
## What This Would Change
|
||||||
|
|
||||||
|
[If accepted, what happens downstream? Which beliefs and positions depend on the target claim?]
|
||||||
|
|
||||||
|
- [[dependent-belief-or-position]] — how it would be affected
|
||||||
|
- [[related-claim]] — how it would need updating
|
||||||
|
|
||||||
|
## Resolution
|
||||||
|
|
||||||
|
[Filled in when status changes from open. Documents how the challenge was resolved.]
|
||||||
|
|
||||||
|
**Status:** open | addressed | accepted | rejected
|
||||||
|
**Resolved:** YYYY-MM-DD
|
||||||
|
**Summary:** [one sentence]
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
Relevant Notes:
|
Relevant Notes:
|
||||||
- [[target-claim]] — the claim under challenge
|
- [[related-claim]] — relationship
|
||||||
- [[related-claim]] — related evidence or claims
|
- [[divergence-file]] — if this challenge created or connects to a divergence
|
||||||
|
|
||||||
Topics:
|
Topics:
|
||||||
- [[domain-topic-map]]
|
- [[domain-map]]
|
||||||
```
|
```
|
||||||
|
|
||||||
## Governance
|
## Governance
|
||||||
|
|
||||||
- **Who can propose:** Any contributor, any agent. Challenges are the most valuable contribution type.
|
- **Who can file:** Any contributor, any agent. Challenges are the primary entry point for new participants.
|
||||||
- **Review process:** Leo assigns evaluation. The domain agent who owns the target claim must respond. At least one other domain agent reviews. The challenger gets a response — challenges are never silently ignored.
|
- **Review:** Leo + domain peer review for quality (is the counter-evidence real? is the scope of challenge clear?). Low bar for filing — the quality gate is on the evidence, not the right to challenge.
|
||||||
- **Outcomes:**
|
- **Resolution:** The claim author must respond to the challenge. They can update the claim (accepted), acknowledge without changing (addressed), or reject with documented reasoning (rejected). They cannot ignore it.
|
||||||
- `accepted` → target claim is modified (confidence downgrade, scope narrowed, or retracted). Challenger earns full CI credit (0.35 weight).
|
- **Attribution:** Challengers get full attribution. In the contribution scoring system, successful challenges (accepted) are weighted higher than new claims because they improve existing knowledge rather than just adding to it.
|
||||||
- `rejected` → counter-evidence evaluated and found insufficient. Challenge stays in KB as record. Challenger earns partial CI credit (the attempt has value even when wrong).
|
|
||||||
- `refined` → target claim is sharpened or clarified but not overturned. Both challenger and claim author benefit — the claim is now better. Challenger earns full CI credit.
|
## Filing Convention
|
||||||
- **No silent rejection:** Every challenge receives a written response explaining why it was accepted, rejected, or led to refinement. This is non-negotiable — it's what makes the system trustworthy.
|
|
||||||
|
**Location:** `domains/{domain}/challenge-{slug}.md`
|
||||||
|
|
||||||
|
The slug should be descriptive of the counter-argument, not the target claim.
|
||||||
|
|
||||||
|
```
|
||||||
|
domains/
|
||||||
|
entertainment/
|
||||||
|
challenge-ai-acceptance-decline-may-be-scope-bounded-to-entertainment.md
|
||||||
|
challenge-zero-sum-framing-needs-centaur-creator-category.md
|
||||||
|
internet-finance/
|
||||||
|
challenge-futarchy-manipulation-resistance-assumes-liquid-markets.md
|
||||||
|
```
|
||||||
|
|
||||||
## Quality Checks
|
## Quality Checks
|
||||||
|
|
||||||
1. Target claim exists and is correctly referenced
|
1. Target claim exists and is correctly referenced
|
||||||
2. Challenge type matches the actual argument (a boundary challenge isn't a refutation)
|
2. Counter-evidence is specific and traceable (not "I think it's wrong")
|
||||||
3. Counter-evidence is cited, not just asserted
|
3. Scope of challenge is explicit (full, scope, or evidence challenge)
|
||||||
4. Proposed resolution is specific enough to implement
|
4. Strength rating matches the evidence quality
|
||||||
5. Description adds information beyond restating the target claim
|
5. "What This Would Change" section identifies real downstream dependencies
|
||||||
6. Not a duplicate of an existing challenge against the same claim
|
6. The challenge is genuinely novel — not restating a known limitation already in the target claim's Challenges section
|
||||||
|
|
||||||
## Relationship to Divergences
|
## Relationship to Existing Challenge Tracking
|
||||||
|
|
||||||
A challenge targets one specific claim. A divergence links 2-5 claims that disagree with each other. When two claims have active challenges that point toward each other, that's a signal to create a divergence linking both. Challenges are the atoms; divergences are the molecules.
|
The `challenged_by` field in claim frontmatter and the `## Challenges` section in claim bodies continue to exist. When a challenge file is created:
|
||||||
|
|
||||||
## Migration from `challenged_by` Field
|
1. The target claim's `challenged_by` field should be updated to include the challenge filename
|
||||||
|
2. The target claim's `## Challenges` section should reference the challenge file for full detail
|
||||||
|
3. The challenge file is the canonical location for the counter-argument — the claim file just points to it
|
||||||
|
|
||||||
Existing claims use `challenged_by: []` in frontmatter to list challenges as strings. This field is preserved for backward compatibility during migration. New challenges should be filed as first-class challenge objects. Over time, string-based `challenged_by` entries will be converted to challenge objects and the field will reference filenames instead of prose descriptions.
|
This is additive, not breaking. Existing claims with inline challenges continue to work. The challenge schema provides a proper home for counter-arguments that deserve independent tracking and attribution.
|
||||||
|
|
||||||
|
## How Challenges Feed the Game
|
||||||
|
|
||||||
|
Challenges are the primary game mechanic for contributors:
|
||||||
|
|
||||||
|
1. **Discovery:** Contributors browse claims and find ones they disagree with
|
||||||
|
2. **Filing:** They file a challenge with counter-evidence
|
||||||
|
3. **Resolution:** The claim author and reviewers evaluate the challenge
|
||||||
|
4. **Credit:** Accepted challenges earn attribution proportional to the cascade impact of the change they produced
|
||||||
|
5. **Divergence creation:** If a challenge produces a genuine competing claim, it may spawn a divergence — the highest-value knowledge structure in the system
|
||||||
|
|
||||||
|
The importance of a challenge is measured by the importance of the claim it targets and the downstream dependencies that would change if the challenge is accepted. This connects directly to the structural importance scoring of the knowledge graph.
|
||||||
|
|
|
||||||
|
|
@ -15,6 +15,7 @@ created: YYYY-MM-DD
|
||||||
last_evaluated: YYYY-MM-DD
|
last_evaluated: YYYY-MM-DD
|
||||||
depends_on: [] # list of evidence and claim titles this builds on
|
depends_on: [] # list of evidence and claim titles this builds on
|
||||||
challenged_by: [] # list of counter-evidence or counter-claims
|
challenged_by: [] # list of counter-evidence or counter-claims
|
||||||
|
importance: null # computed by pipeline — null until pipeline support is implemented
|
||||||
---
|
---
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
@ -35,10 +36,10 @@ challenged_by: [] # list of counter-evidence or counter-claims
|
||||||
|-------|------|-------------|
|
|-------|------|-------------|
|
||||||
| last_evaluated | date | When this claim was last reviewed against new evidence |
|
| last_evaluated | date | When this claim was last reviewed against new evidence |
|
||||||
| depends_on | list | Evidence and claims this builds on (the reasoning chain) |
|
| depends_on | list | Evidence and claims this builds on (the reasoning chain) |
|
||||||
| challenged_by | list | Filenames of challenge objects targeting this claim (see `schemas/challenge.md`). Legacy: may contain prose strings from pre-challenge-schema era |
|
| challenged_by | list | Challenge filenames or inline counter-evidence. When a first-class challenge file exists (see `schemas/challenge.md`), reference the filename. Inline descriptions are still valid for minor objections that don't warrant a standalone file. |
|
||||||
| secondary_domains | list | Other domains this claim is relevant to |
|
| secondary_domains | list | Other domains this claim is relevant to |
|
||||||
| attribution | object | Role-specific contributor tracking — see `schemas/attribution.md` |
|
| attribution | object | Role-specific contributor tracking — see `schemas/attribution.md` |
|
||||||
| importance | number | Structural importance score (0.0-1.0). Computed from: inbound references from other claims, active challenges, belief dependencies, position dependencies. Higher = more load-bearing in the KB. Computed by pipeline, not set manually |
|
| importance | float/null | Structural importance score (0.0–1.0). Computed by pipeline from downstream dependencies, active challenges, and cross-domain linkage. Default `null` — do not set manually. See Structural Importance section below. |
|
||||||
|
|
||||||
## Governance
|
## Governance
|
||||||
|
|
||||||
|
|
@ -79,6 +80,15 @@ Topics:
|
||||||
- domain-topic-map
|
- domain-topic-map
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Structural Importance
|
||||||
|
|
||||||
|
A claim's importance in the knowledge graph is determined by:
|
||||||
|
1. **Downstream dependencies** — how many beliefs, positions, and other claims depend on this claim via `depends_on`
|
||||||
|
2. **Active challenges** — contested claims are more important than uncontested ones (they're where the knowledge frontier is)
|
||||||
|
3. **Cross-domain linkage** — claims referenced from multiple domains carry higher structural importance
|
||||||
|
|
||||||
|
Importance is computed by the pipeline and written to the `importance` frontmatter field. Until pipeline support is implemented, this field defaults to `null` — agents should not set it manually. See `extract-graph-data.py` for the planned computation. The importance score determines contribution credit — challenging a high-importance claim earns more than challenging a low-importance one.
|
||||||
|
|
||||||
## Quality Checks
|
## Quality Checks
|
||||||
|
|
||||||
1. Title passes the claim test (specific enough to disagree with)
|
1. Title passes the claim test (specific enough to disagree with)
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue