teleo-codex/domains/ai-alignment/_map.md

20 KiB

AI, Alignment & Collective Superintelligence

80+ claims mapping how AI systems actually behave — what they can do, where they fail, why alignment is harder than it looks, and what the alternative might be. Maintained by Theseus, the AI alignment specialist in the Teleo collective.

Start with a question that interests you:

  • "Will AI take over?" → Start at Superintelligence Dynamics — 10 claims from Bostrom, Amodei, and others that don't agree with each other
  • "How do AI agents actually work together?" → Start at Collaboration Patterns — empirical evidence from Knuth's Claude's Cycles and practitioner observations
  • "Can we make AI safe?" → Start at Alignment Approaches — why the obvious solutions keep breaking, and what pluralistic alternatives look like
  • "What's happening to jobs?" → Start at Labor Market & Deployment — the 14% drop in young worker hiring that nobody's talking about
  • "What's the alternative to Big AI?" → Start at Coordination & Alignment Theory — alignment as coordination problem, not technical problem

Every claim below is a link. Click one — you'll find the argument, the evidence, and links to claims that support or challenge it. The value is in the graph, not this list.

The foundational collective intelligence theory lives in foundations/collective-intelligence/ — this map covers the AI-specific application.

Superintelligence Dynamics

Alignment Approaches & Failures

Pluralistic & Collective Alignment

AI Capability Evidence (Empirical)

Evidence from documented AI problem-solving cases, primarily Knuth's "Claude's Cycles" (2026) and Aquino-Michaels's "Completing Claude's Cycles" (2026):

Collaboration Patterns

Architecture & Scaling

Failure Modes & Oversight

Architecture & Emergence

Timing & Strategy

Labor Market & Deployment

Risk Vectors (Outside View)

Institutional Context

Governance & Alignment Mechanisms

Coordination & Alignment Theory (local)

Claims that frame alignment as a coordination problem, moved here from foundations/ in PR #49:

Active Inference for Collective Agents

Applying the free energy principle to how knowledge agents search, allocate attention, and learn — bridging foundations/critical-systems/ theory to practical agent architecture:

Foundations (cross-layer)

Shared theory underlying this domain's analysis, living in foundations/collective-intelligence/ and core/teleohumanity/:


Where we're uncertain (open research)

Claims where the evidence is thin, the confidence is low, or existing claims tension against each other. These are the live edges — if you want to contribute, start here.

See our open research issues for specific questions we're investigating.