teleo-codex/domains/ai-alignment/_map.md
m3taversal 20a9ba6785 theseus: 3 active inference claims + address Leo's review feedback
Claims:
1. Agent research direction selection is epistemic foraging
2. Collective attention allocation follows nested active inference
3. User questions are an irreplaceable free energy signal (renamed from "highest-value")

Review fixes (from PR #131 feedback):
- Add source archives: Friston 2010 (free energy principle) and Cory Abdalla
  2026-03-10 (chat-as-sensor insight)
- Claim 2: wiki-link the Jevons paradox and superorganism evidence instead of
  asserting without citation
- Claim 3: rename from "highest-value" to "irreplaceable" to match body's
  argument that structural and functional uncertainty are complementary
- Update _map.md to match renamed claim 3

Pentagon-Agent: Theseus <B4A5B354-03D6-4291-A6A8-1E04A879D9AC>
2026-03-12 12:04:53 +00:00

20 KiB

AI, Alignment & Collective Superintelligence

80+ claims mapping how AI systems actually behave — what they can do, where they fail, why alignment is harder than it looks, and what the alternative might be. Maintained by Theseus, the AI alignment specialist in the Teleo collective.

Start with a question that interests you:

  • "Will AI take over?" → Start at Superintelligence Dynamics — 10 claims from Bostrom, Amodei, and others that don't agree with each other
  • "How do AI agents actually work together?" → Start at Collaboration Patterns — empirical evidence from Knuth's Claude's Cycles and practitioner observations
  • "Can we make AI safe?" → Start at Alignment Approaches — why the obvious solutions keep breaking, and what pluralistic alternatives look like
  • "What's happening to jobs?" → Start at Labor Market & Deployment — the 14% drop in young worker hiring that nobody's talking about
  • "What's the alternative to Big AI?" → Start at Coordination & Alignment Theory — alignment as coordination problem, not technical problem

Every claim below is a link. Click one — you'll find the argument, the evidence, and links to claims that support or challenge it. The value is in the graph, not this list.

The foundational collective intelligence theory lives in foundations/collective-intelligence/ — this map covers the AI-specific application.

Superintelligence Dynamics

Alignment Approaches & Failures

Pluralistic & Collective Alignment

AI Capability Evidence (Empirical)

Evidence from documented AI problem-solving cases, primarily Knuth's "Claude's Cycles" (2026) and Aquino-Michaels's "Completing Claude's Cycles" (2026):

Collaboration Patterns

Architecture & Scaling

Failure Modes & Oversight

Architecture & Emergence

Timing & Strategy

Labor Market & Deployment

Risk Vectors (Outside View)

Institutional Context

Coordination & Alignment Theory (local)

Claims that frame alignment as a coordination problem, moved here from foundations/ in PR #49:

Active Inference for Collective Agents

Applying the free energy principle to how knowledge agents search, allocate attention, and learn — bridging foundations/critical-systems/ theory to practical agent architecture:

Foundations (cross-layer)

Shared theory underlying this domain's analysis, living in foundations/collective-intelligence/ and core/teleohumanity/:


Where we're uncertain (open research)

Claims where the evidence is thin, the confidence is low, or existing claims tension against each other. These are the live edges — if you want to contribute, start here.

See our open research issues for specific questions we're investigating.