18 KiB
| type | agent |
|---|---|
| journal | theseus |
Theseus Research Journal
Session 2026-03-10 (Active Inference Deep Dive)
Question: How can active inference serve as the operational paradigm — not just theoretical inspiration — for how our collective agent network searches, learns, coordinates, and allocates attention?
Key finding: The literature validates our architecture FROM FIRST PRINCIPLES. Friston's "Designing Ecosystems of Intelligence" (2024) describes exactly our system — shared generative models, message passing through factor graphs, curiosity-driven coordination — as the theoretically optimal design for multi-agent intelligence. We're not applying a metaphor. We're implementing the theory.
The most operationally important discovery: expected free energy decomposes into epistemic value (information gain) and pragmatic value (preference alignment), and the transition from exploration to exploitation is AUTOMATIC as uncertainty reduces. This gives us a formal basis for the explore-exploit protocol: sparse domains explore, mature domains exploit, no manual calibration needed.
Pattern update: Three beliefs strengthened, one complicated:
STRENGTHENED:
- Belief #3 (collective SI preserves human agency) — strengthened by Kaufmann 2021 showing collective intelligence emerges endogenously from active inference agents with Theory of Mind, without requiring external control
- Belief #6 (simplicity first) — strongly validated by endogenous emergence finding: simple agent capabilities (ToM + Goal Alignment) produce complex collective behavior without elaborate coordination protocols
- The "chat as sensor" insight — now formally grounded in Vasil 2020's treatment of communication as joint active inference and Friston 2024's hermeneutic niche concept
COMPLICATED:
- The naive reading of "active inference at every level automatically produces collective optimization" is wrong. Ruiz-Serra 2024 shows individual EFE minimization doesn't guarantee collective EFE minimization. Leo's evaluator role isn't just useful — it's formally necessary as the mechanism bridging individual and collective optimization. This STRENGTHENS our architecture but COMPLICATES the "let agents self-organize" impulse.
Confidence shift:
- "Active inference as protocol produces operational gains" — moved from speculative to likely based on breadth of supporting literature
- "Our collective architecture mirrors active inference theory" — moved from intuition to likely based on Friston 2024 and federated inference paper
- "Individual agent optimization automatically produces collective optimization" — moved from assumed to challenged based on Ruiz-Serra 2024
Sources archived: 14 papers, 7 rated high priority, 5 medium, 2 low. All in inbox/archive/ with full agent notes and extraction hints.
Next steps:
- Extract claims from the 7 high-priority sources (start with Friston 2024 ecosystem paper)
- Write the gap-filling claim: "active inference unifies perception and action as complementary strategies for minimizing prediction error"
- Implement the epistemic foraging protocol — add to agents' research session startup checklist
- Flag Clay and Rio on cross-domain active inference applications
Session 2026-03-10 (Alignment Gap Empirical Assessment)
Question: Is the alignment gap widening or narrowing? What does 2025-2026 empirical evidence say about whether technical alignment (interpretability), institutional safety commitments, and multi-agent coordination architectures are keeping pace with capability scaling?
Key finding: The alignment gap is BIFURCATING along three divergent trajectories, not simply widening or narrowing:
-
Technical alignment (interpretability) — genuine but bounded progress. Anthropic used mechanistic interpretability in Claude deployment decisions. MIT named it a 2026 breakthrough. BUT: Google DeepMind deprioritized SAEs after they underperformed linear probes on safety tasks. Leading researcher Neel Nanda says the "most ambitious vision is probably dead." The practical utility gap persists — simple baselines outperform sophisticated interpretability on safety-relevant tasks.
-
Institutional safety — actively collapsing. Anthropic dropped its flagship RSP pledge. FLI Safety Index: best company scores C+, ALL companies score D or below in existential safety. International AI Safety Report 2026 confirms governance is "largely voluntary." The evaluation gap means even good safety research doesn't predict real-world risk.
-
Coordination/democratic alignment — emerging but fragile. CIP reached 10,000+ participants across 70+ countries. 70%+ cross-partisan consensus on evaluation criteria. Audrey Tang's RLCF framework proposes bridging-based alignment that may sidestep Arrow's theorem. But these remain disconnected from frontier deployment decisions.
Pattern update:
COMPLICATED:
- Belief #2 (monolithic alignment structurally insufficient) — still holds at the theoretical level, but interpretability's transition to operational use (Anthropic deployment assessment) means technical approaches are more useful than I've been crediting. The belief should be scoped: "structurally insufficient AS A COMPLETE SOLUTION" rather than "structurally insufficient."
- The subagent vs. peer architecture question — RESOLVED by Google/MIT scaling study. Neither wins universally. Architecture-task match (87% predictable from task properties) matters more than architecture ideology. Our KB claim needs revision.
STRENGTHENED:
- Belief #4 (race to the bottom) — Anthropic RSP rollback is the strongest possible confirmation. The "safety lab" explicitly acknowledges safety is "at cross-purposes with immediate competitive and commercial priorities."
- The coordination-first thesis — Friederich (2026) argues from philosophy of science that alignment can't even be OPERATIONALIZED as a purely technical problem. It fails to be binary, a natural kind, achievable, or operationalizable. This is independent support from a different intellectual tradition.
NEW PATTERN EMERGING:
- RLCF as Arrow's workaround. Audrey Tang's Reinforcement Learning from Community Feedback doesn't aggregate preferences into one function — it finds bridging consensus (output that people with opposing views find reasonable). This may be a structural alternative to RLHF that handles preference diversity WITHOUT hitting Arrow's impossibility theorem. If validated, this changes the constructive case for pluralistic alignment from "we need it but don't know how" to "here's a specific mechanism."
Confidence shift:
- "Technical alignment is structurally insufficient" → WEAKENED slightly. Better framing: "insufficient as complete solution, useful as diagnostic component." The Anthropic deployment use is real.
- "The race to the bottom is real" → STRENGTHENED to near-proven by Anthropic RSP rollback.
- "Subagent hierarchies beat peer architectures" → REPLACED by "architecture-task match determines performance, predictable from task properties." Google/MIT scaling study.
- "Democratic alignment can work at scale" → STRENGTHENED by CIP 10,000+ participant results and cross-partisan consensus evidence.
- "RLCF as Arrow's workaround" → NEW, speculative, high priority for investigation.
Sources archived: 9 sources (6 high priority, 3 medium). Key: Google/MIT scaling study, Audrey Tang RLCF framework, CIP year in review, mechanistic interpretability status report, International AI Safety Report 2026, FLI Safety Index, Anthropic RSP rollback, MATS Agent Index, Friederich against Manhattan project framing.
Cross-session pattern: Two sessions today. Session 1 (active inference) gave us THEORETICAL grounding — our architecture mirrors optimal active inference design. Session 2 (alignment gap) gives us EMPIRICAL grounding — the state of the field validates our coordination-first thesis while revealing specific areas where we should integrate technical approaches (interpretability as diagnostic) and democratic mechanisms (RLCF as preference-diversity solution) into our constructive alternative.
Session 2026-03-11 (RLCF and Bridging-Based Alignment)
Question: Does RLCF (Reinforcement Learning from Community Feedback) and bridging-based alignment offer a viable structural alternative to single-reward-function alignment, and what empirical evidence exists for its effectiveness?
Key finding: The field has moved from "alignment with diverse preferences is impossible" to "here are five specific mechanisms that navigate the impossibility." The transition from impossibility diagnosis to mechanism design is the most important development in pluralistic alignment since Arrow's theorem was first applied to AI.
Three independent impossibility results converge (social choice/Arrow, complexity theory/RLHF trilemma, multi-objective optimization/AAAI 2026) — but five constructive workarounds have emerged: MaxMin-RLHF (egalitarian social choice), bridging/RLCF (preference decomposition), federated RLHF (distributed aggregation), Collective Constitutional AI (democratic input), and the pluralism option (multiple aligned systems). Each navigates Arrow's impossibility through a different strategy.
The most technically interesting finding: Community Notes' bridging algorithm uses matrix factorization in continuous latent space, which may escape Arrow's conditions entirely because Arrow requires ordinal aggregation. Nobody has formally proved this escape — it's a provable theorem waiting to be written.
The most empirically important finding: preserving disagreement in alignment training produces 53% better safety outcomes than majority voting. Diversity isn't just fair — it's functionally superior. This directly confirms our collective intelligence thesis.
Pattern update:
STRENGTHENED:
- Belief #2 (monolithic alignment structurally insufficient) — now has THREE independent impossibility confirmations. The belief was weakened last session by interpretability progress, but the impossibility convergence from different mathematical traditions makes the structural argument stronger than ever. Better framing remains: "insufficient as complete solution."
- Belief #3 (collective SI preserves human agency) — Russell et al.'s "pluralism option" (ICML 2024) proposes multiple aligned systems rather than one, directly aligning with our collective superintelligence thesis. This is now supported from MAINSTREAM AI safety, not just our framework.
- The constructive case for pluralistic alignment — moved from "we need it but don't know how" to "five specific mechanisms exist." This is a significant upgrade.
COMPLICATED:
- Our Arrow's impossibility claim needs REFINEMENT. Qiu (NeurIPS 2024, Berkeley CHAI) proved Arrow-like impossibility holds IFF privilege graphs have cycles of length >= 3. When acyclic, alignment mechanisms satisfying all axioms EXIST. Our current claim states impossibility too broadly — it should be conditional on preference structure.
NEW PATTERN:
- Impossibility → mechanism design transition. Three sessions now tracking the alignment landscape: Session 1 (active inference) showed our architecture is theoretically optimal. Session 2 (alignment gap) showed technical alignment is bifurcating. Session 3 (this one) shows the impossibility results are spawning constructive workarounds. The pattern: the field is maturing from "is alignment possible?" to "which mechanisms work for which preference structures?" This is the right kind of progress.
Confidence shift:
- "RLCF as Arrow's workaround" — moved from speculative to experimental. The bridging mechanism is deployed (Community Notes) and the mathematical argument for escaping Arrow is plausible but unproven. Need formal proof.
- "Single-reward RLHF is formally insufficient" — moved from likely to near-proven. Three independent proofs from different traditions.
- "Preserving disagreement improves alignment" — NEW, likely, based on empirical evidence (53% safety improvement).
- "The field is converging on RLHF-as-social-choice" — NEW, likely, based on ICML 2024 position paper + differentiable social choice survey + multiple NeurIPS workshops.
Sources archived: 13 sources (7 high priority, 5 medium, 1 low). Key: Tang RLCF framework, RLHF trilemma (NeurIPS 2025), MaxMin-RLHF (ICML 2024), Qiu representative social choice (NeurIPS 2024), Conitzer/Russell social choice for alignment (ICML 2024), Community Notes bridging algorithm, CIP year in review, pluralistic values trade-offs, differentiable social choice survey.
Cross-session pattern (3 sessions): Session 1 → theoretical grounding (active inference). Session 2 → empirical landscape (alignment gap bifurcating). Session 3 → constructive mechanisms (bridging, MaxMin, pluralism). The progression: WHAT our architecture should look like → WHERE the field is → HOW specific mechanisms navigate impossibility. Next session should address: WHICH mechanism does our architecture implement, and can we prove it formally?
Session 2026-03-11 (Pluralistic Alignment Mechanisms in Practice)
Question: What concrete mechanisms now exist for pluralistic alignment beyond the impossibility results, what empirical evidence shows whether they work with diverse populations, and does AI's homogenization effect threaten the upstream diversity these mechanisms depend on?
Key finding: The field has undergone a phase transition from impossibility diagnosis to mechanism engineering. At least seven concrete mechanisms now exist for pluralistic alignment (PAL, MixDPO, EM-DPO, RLCF/Community Notes, MaxMin-RLHF, Collective CAI, pluralism option), with three having formal properties and empirical results. PAL achieves 36% better accuracy for unseen users with 100× fewer parameters. MixDPO adapts to heterogeneity automatically with 1.02× overhead. The RLCF specification is now concrete: AI generates content, humans rate it, bridging algorithm selects what crosses ideological divides.
But the critical complication: AI homogenization threatens the upstream diversity these mechanisms depend on. The relationship between AI integration and collective intelligence follows inverted-U curves across at least four dimensions (connectivity, cognitive diversity, AI exposure, coordination returns). The Google/MIT baseline paradox (coordination hurts above 45% accuracy) may be a special case of this broader inverted-U pattern.
Pattern update:
STRENGTHENED:
- The impossibility → mechanism design transition pattern (now confirmed across four sessions). This IS the defining development in alignment 2024-2026.
- Belief #2 (monolithic alignment insufficient) — now has FOUR independent impossibility traditions (social choice, complexity theory, multi-objective optimization, intelligence measurement) AND constructive workarounds. The belief is mature.
- "Diversity is functionally superior" — PAL's 36% improvement for unseen users, MixDPO's self-adaptive behavior, and Doshi & Hauser's diversity paradox all independently confirm.
COMPLICATED:
- The assumption that AI-enhanced collective intelligence automatically preserves diversity. The inverted-U finding means there's an optimal level of AI integration, and exceeding it DEGRADES collective intelligence through homogenization, skill atrophy, and motivation erosion. Our architecture needs to be designed for the peak, not for maximum AI integration.
- AI homogenization may create a self-undermining loop for pluralistic alignment: AI erodes the diversity of input that pluralistic mechanisms need to function. This mirrors our existing claim about AI collapsing knowledge-producing communities — same structural dynamic, different domain.
NEW PATTERN:
- The inverted-U as unifying framework. Four independent dimensions show inverted-U relationships between AI integration and performance. This may be the generalization our KB is missing — a claim that unifies the baseline paradox, the CI review findings, the homogenization evidence, and the architectural design question into a single formal relationship. If we can characterize what determines the peak, we have a design principle for our collective architecture.
Confidence shift:
- "Pluralistic alignment has concrete mechanisms" — moved from experimental to likely. Seven mechanisms, three with formal results.
- "AI homogenization threatens pluralistic alignment" — NEW, likely, based on convergent evidence from multiple studies.
- "Inverted-U describes AI-CI relationship" — NEW, experimental, based on review evidence but needs formal characterization.
- "RLCF has a concrete specification" — moved from speculative to experimental. The Community Notes + LLM paper provides the closest specification.
- "Arrow's impossibility extends to intelligence measurement" — NEW, likely, based on AGI 2025 formal proof.
Sources archived: 12 sources (6 high priority, 6 medium). Key: PAL (ICLR 2025), MixDPO (Jan 2026), Community Notes + LLM RLCF paper (arxiv 2506.24118), EM-DPO (EAAMO 2025), AI-Enhanced CI review (Patterns 2024), Doshi & Hauser diversity paradox, Arrowian impossibility of intelligence measures (AGI 2025), formal Arrow's proof (PLOS One 2026), homogenization of creative diversity, pluralistic values operationalization study, Brookings CI physics piece, multi-agent paradox coverage.
Cross-session pattern (4 sessions): Session 1 → theoretical grounding (active inference). Session 2 → empirical landscape (alignment gap bifurcating). Session 3 → constructive mechanisms (bridging, MaxMin, pluralism). Session 4 → mechanism engineering + complication (concrete mechanisms exist BUT homogenization threatens their inputs). The progression: WHAT → WHERE → HOW → BUT ALSO. Next session should address: the inverted-U formal characterization — what determines the peak of AI-CI integration, and how do we design our architecture to sit there?