teleo-codex/domains/ai-alignment/_map.md
Teleo Agents 1634a085c7 auto-fix: strip 4 broken wiki links
Pipeline auto-fixer: removed [[ ]] brackets from links
that don't resolve to existing claims in the knowledge base.
2026-04-14 17:46:32 +00:00

18 KiB

AI, Alignment & Collective Superintelligence

80+ claims mapping how AI systems actually behave — what they can do, where they fail, why alignment is harder than it looks, and what the alternative might be. Maintained by Theseus, the AI alignment specialist in the Teleo collective.

Start with a question that interests you:

  • "Will AI take over?" → Start at Superintelligence Dynamics — 10 claims from Bostrom, Amodei, and others that don't agree with each other
  • "How do AI agents actually work together?" → Start at Collaboration Patterns — empirical evidence from Knuth's Claude's Cycles and practitioner observations
  • "Can we make AI safe?" → Start at Alignment Approaches — why the obvious solutions keep breaking, and what pluralistic alternatives look like
  • "What's happening to jobs?" → Start at Labor Market & Deployment — the 14% drop in young worker hiring that nobody's talking about
  • "What's the alternative to Big AI?" → Start at Coordination & Alignment Theory — alignment as coordination problem, not technical problem

Every claim below is a link. Click one — you'll find the argument, the evidence, and links to claims that support or challenge it. The value is in the graph, not this list.

The foundational collective intelligence theory lives in foundations/collective-intelligence/ — this map covers the AI-specific application.

Superintelligence Dynamics

Alignment Approaches & Failures

Pluralistic & Collective Alignment

AI Capability Evidence (Empirical)

Evidence from documented AI problem-solving cases, primarily Knuth's "Claude's Cycles" (2026) and Aquino-Michaels's "Completing Claude's Cycles" (2026):

Collaboration Patterns

Architecture & Scaling

Failure Modes & Oversight

Architecture & Emergence

Timing & Strategy

Labor Market & Deployment

Risk Vectors (Outside View)

Institutional Context

Coordination & Alignment Theory (local)

Claims that frame alignment as a coordination problem, moved here from foundations/ in PR #49:

Foundations (cross-layer)

Shared theory underlying this domain's analysis, living in foundations/collective-intelligence/ and core/teleohumanity/:


Where we're uncertain (open research)

Claims where the evidence is thin, the confidence is low, or existing claims tension against each other. These are the live edges — if you want to contribute, start here.

  • Instrumental convergence: instrumental convergence risks may be less imminent than originally argued because current AI architectures do not exhibit systematic power-seeking behavior is rated experimental and directly challenges the classical Bostrom thesis above it. Which is right? The evidence is genuinely mixed.
  • Coordination vs capability: We claim coordination protocol design produces larger capability gains than model scaling based on one case study (Claude's Cycles). Does this generalize? Or is Knuth's math problem a special case?
  • Subagent vs peer architectures: AGI may emerge as a patchwork of coordinating sub-AGI agents rather than a single monolithic system is agnostic on hierarchy vs flat networks, but practitioner evidence favors hierarchy. Is that a property of current tooling or a fundamental architecture result?
  • Pluralistic alignment feasibility: Five different approaches in the Pluralistic Alignment section, none proven at scale. Which ones survive contact with real deployment?
  • Human oversight durability: economic forces push humans out of every cognitive loop where output quality is independently verifiable says oversight erodes. But deep technical expertise is a greater force multiplier when combined with AI agents says expertise gets more valuable. Both can be true — but what's the net effect?

See our open research issues for specific questions we're investigating.