Compare commits

..

1 commit

Author SHA1 Message Date
Teleo Agents
7b0329c050 theseus: extract from 2024-11-00-ruiz-serra-factorised-active-inference-multi-agent.md
- Source: inbox/archive/2024-11-00-ruiz-serra-factorised-active-inference-multi-agent.md
- Domain: ai-alignment
- Extracted by: headless extraction cron (worker 6)

Pentagon-Agent: Theseus <HEADLESS>
2026-03-12 13:13:08 +00:00
7 changed files with 57 additions and 77 deletions

View file

@ -25,7 +25,7 @@ Since [[the internet enabled global communication but not global cognition]], th
### Additional Evidence (confirm)
*Source: [[2024-11-00-ruiz-serra-factorised-active-inference-multi-agent]] | Added: 2026-03-12 | Extractor: anthropic/claude-sonnet-4.5*
Ruiz-Serra et al. (2024) provide formal evidence that individual free energy minimization in multi-agent active inference systems does not guarantee collective free energy minimization. The ensemble-level expected free energy characterizes basins of attraction that may not align with individual optima. This demonstrates mathematically that even when each agent is individually rational and aligned, coordination problems persist—the interaction structure itself determines whether individual optimization produces collective intelligence or collective failure. The finding validates that alignment cannot be solved at the individual agent level alone; explicit coordination mechanisms are structurally necessary.
Ruiz-Serra et al. (2024) provide formal evidence that individual free energy minimization in multi-agent active inference systems does not guarantee collective free energy minimization. The ensemble-level expected free energy 'characterizes basins of attraction of games with multiple Nash Equilibria under different conditions' but 'is not necessarily minimised at the aggregate level.' This demonstrates mathematically that alignment cannot be solved at the individual agent level—coordination mechanisms are structurally necessary to bridge individual and collective optimization. Testing on 2-3 player iterated normal-form games shows that interaction structure (game form, communication channels) determines whether individual rationality produces collective benefit.
---

View file

@ -37,12 +37,6 @@ The finding also strengthens [[no research group is building alignment through c
Since [[the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it]], coordination-based alignment that *increases* capability rather than taxing it would face no race-to-the-bottom pressure. The Residue prompt is alignment infrastructure that happens to make the system more capable, not less.
### Additional Evidence (confirm)
*Source: [[2024-11-00-ruiz-serra-factorised-active-inference-multi-agent]] | Added: 2026-03-12 | Extractor: anthropic/claude-sonnet-4.5*
Ruiz-Serra et al. demonstrate that the interaction structure (game form, communication channels, coordination mechanisms) determines whether individual agent optimization produces collective optimization. The same agents with the same individual capabilities produce radically different collective outcomes depending on the coordination protocol. This provides formal game-theoretic evidence that protocol design—not agent capability—is the primary determinant of multi-agent system performance. Individual free energy minimization is insufficient; the coordination structure shapes the basins of attraction that determine collective behavior.
---
Relevant Notes:

View file

@ -0,0 +1,35 @@
---
type: claim
domain: ai-alignment
secondary_domains: [collective-intelligence]
description: "Factorised generative models where agents maintain explicit beliefs about other agents' internal states enable strategic coordination without centralized control"
confidence: experimental
source: "Ruiz-Serra et al., 'Factorised Active Inference for Strategic Multi-Agent Interactions' (AAMAS 2025)"
created: 2026-03-11
---
# Factorised generative models enable decentralized multi-agent coordination through individual-level beliefs about other agents' internal states
Ruiz-Serra et al. introduce a factorisation approach where each active inference agent maintains "explicit, individual-level beliefs about the internal states of other agents" rather than relying on centralized coordination or shared world models. This factorisation enables agents to perform "strategic planning in a joint context" by modeling other agents' beliefs, preferences, and likely actions—essentially implementing Theory of Mind within the active inference framework.
The key architectural innovation is that agents don't need access to other agents' actual internal states or a global coordinator. Instead, each agent constructs and updates its own beliefs about what other agents believe and will do, using these beliefs for strategic action selection. This decentralized approach scales better than centralized coordination while enabling more sophisticated strategic interaction than agents without Theory of Mind capabilities.
The framework was tested on iterated normal-form games with 2-3 players, demonstrating that agents can navigate both cooperative and competitive strategic interactions using only their individual beliefs about others. This provides a computational implementation of Theory of Mind that could be applied to multi-agent AI systems requiring strategic coordination without centralized control.
## Evidence
- Ruiz-Serra et al. (2024) formalize factorised generative models where each agent maintains individual-level beliefs about other agents' internal states
- The framework enables strategic planning in joint contexts without requiring centralized coordination or shared world models
- Validation through 2-3 player iterated normal-form games shows agents can handle cooperative and non-cooperative strategic interactions using factorised beliefs
- The approach operationalizes Theory of Mind within active inference by having agents model other agents' beliefs and preferences
## Architectural Implications
This factorisation approach provides a middle path between fully independent agents (no coordination) and centrally coordinated systems (single point of failure). For multi-agent AI architectures, it suggests that giving agents the capacity to model each other's internal states enables strategic coordination without requiring a central controller or shared knowledge base. However, as shown in the companion claim on individual-collective optimization tensions, this capability alone does not guarantee collectively optimal outcomes—interaction structure design remains critical.
---
Relevant Notes:
- [[AI agent orchestration that routes data and tools between specialized models outperforms both single-model and human-coached approaches because the orchestrator contributes coordination not direction]]
- [[collective intelligence requires diversity as a structural precondition not a moral preference]]
- [[individual-free-energy-minimization-does-not-guarantee-collective-optimization-in-multi-agent-active-inference]]

View file

@ -1,46 +0,0 @@
---
type: claim
domain: ai-alignment
description: "Agents maintain explicit individual-level beliefs about other agents' internal states through model factorisation, enabling strategic planning without centralized coordination"
confidence: experimental
source: "Ruiz-Serra et al., 'Factorised Active Inference for Strategic Multi-Agent Interactions' (AAMAS 2025)"
created: 2026-03-11
secondary_domains: [collective-intelligence]
---
# Factorised generative models enable decentralized Theory of Mind in multi-agent active inference systems
Active inference agents can maintain explicit, individual-level beliefs about the internal states of other agents through factorisation of the generative model. This enables each agent to perform strategic planning in a joint context without requiring centralized coordination or a global model of the system.
The factorisation approach operationalizes Theory of Mind within the active inference framework: each agent models not just the observable behavior of others, but their internal states—beliefs, preferences, and decision-making processes. This allows agents to anticipate others' actions based on inferred mental states rather than just observed patterns.
## Evidence
Ruiz-Serra et al. (2024) demonstrate this through:
1. **Factorised generative models**: Each agent maintains a separate model component for each other agent's internal state
2. **Strategic planning**: Agents use these beliefs about others' internal states for planning in iterated normal-form games
3. **Decentralized representation**: The multi-agent system is represented in a decentralized way—no agent needs a global view
4. **Game-theoretic validation**: The framework successfully navigates cooperative and non-cooperative strategic interactions in 2- and 3-player games
## Implications
This architecture provides a computational implementation of Theory of Mind that:
- Scales to multi-agent systems without centralized coordination
- Enables strategic reasoning about others' likely actions based on inferred beliefs
- Maintains individual agent autonomy while supporting coordination
- Provides a formal framework for modeling how agents model each other
The approach bridges active inference (a normative theory of intelligent behavior) with game theory (a normative theory of strategic interaction).
---
Relevant Notes:
- [[AI alignment is a coordination problem not a technical problem]]
- [[intelligence is a property of networks not individuals]]
- [[coordination protocol design produces larger capability gains than model scaling because the same AI model performed 6x better with structured exploration than with human coaching on the same problem]]
Topics:
- [[domains/ai-alignment/_map]]
- [[foundations/collective-intelligence/_map]]

View file

@ -1,44 +1,35 @@
---
type: claim
domain: ai-alignment
description: "Individual free energy minimization in multi-agent active inference does not guarantee collective free energy minimization because ensemble-level expected free energy characterizes basins of attraction that may not align with individual optima"
secondary_domains: [collective-intelligence]
description: "Individual free energy minimization in multi-agent active inference does not guarantee collective free energy minimization; interaction structure determines collective outcomes"
confidence: experimental
source: "Ruiz-Serra et al., 'Factorised Active Inference for Strategic Multi-Agent Interactions' (AAMAS 2025)"
created: 2026-03-11
secondary_domains: [collective-intelligence]
---
# Individual free energy minimization does not guarantee collective optimization in multi-agent active inference systems
When multiple active inference agents interact strategically, each agent minimizes its own expected free energy (EFE) based on beliefs about other agents' internal states. However, the ensemble-level expected free energy—which characterizes basins of attraction in games with multiple Nash Equilibria—is not necessarily minimized at the aggregate level.
When multiple active inference agents interact strategically, each agent minimizing its individual expected free energy (EFE) does not necessarily produce optimal collective outcomes. Ruiz-Serra et al. demonstrate through game-theoretic analysis that "ensemble-level expected free energy characterizes basins of attraction of games with multiple Nash Equilibria under different conditions" but "it is not necessarily minimised at the aggregate level."
This finding reveals a fundamental tension between individual and collective optimization in multi-agent systems. Even when each agent is individually rational and minimizing its own free energy, the collective outcome can be suboptimal. The specific interaction structure (game type, communication channels, coordination mechanisms) determines whether individual optimization produces collective intelligence or collective failure.
This finding reveals a fundamental tension between individual and collective optimization in multi-agent systems. Each agent operating under active inference principles will minimize its own free energy through belief updating and action selection, but the interaction structure (game form, communication channels, coordination mechanisms) determines whether these individual optimizations produce collectively beneficial outcomes.
The paper applies factorised generative models where each agent maintains "explicit, individual-level beliefs about the internal states of other agents" to enable strategic planning in joint contexts—essentially operationalizing Theory of Mind within active inference. Testing this framework on iterated normal-form games with 2-3 players shows that while agents can navigate cooperative and non-cooperative strategic interactions, the aggregate system behavior depends critically on the specific game structure.
## Evidence
Ruiz-Serra et al. (2024) demonstrate this through factorised generative models where each agent maintains explicit individual-level beliefs about other agents' internal states. In iterated normal-form games with 2 and 3 players, they show that:
1. Agents successfully use beliefs about others' internal states for strategic planning (operationalizing Theory of Mind within active inference)
2. The ensemble-level EFE characterizes basins of attraction under different conditions
3. Individual free energy minimization does not guarantee that ensemble-level EFE is minimized
This is not a failure of the framework but a feature: multi-agent systems have genuine coordination problems that cannot be solved by individual rationality alone.
- Ruiz-Serra et al. (2024) show through formal analysis that ensemble-level EFE characterizes equilibrium basins but is not necessarily minimized at aggregate level in multi-agent active inference systems
- Game-theoretic simulations with 2-3 player iterated normal-form games demonstrate that individual free energy minimization can produce suboptimal collective outcomes depending on interaction structure
- Factorised generative models enable agents to maintain individual-level beliefs about other agents' internal states, but this Theory of Mind capability does not automatically resolve individual-collective optimization tensions
## Implications
This result has direct architectural implications for AI agent systems:
- **Explicit coordination mechanisms are necessary**: Simply giving each agent active inference dynamics and assuming collective optimization is insufficient
- **Interaction structure design matters**: The form of agent interaction (review processes, communication protocols, cross-domain synthesis) shapes whether individual research produces collective intelligence
- **Evaluator roles are formally justified**: Roles like cross-domain synthesis exist precisely because individual agent optimization doesn't guarantee collective optimization
This result has direct architectural implications for multi-agent AI systems. It means that giving each agent active inference dynamics and assuming collective intelligence will emerge is insufficient—explicit coordination mechanisms are required to bridge individual and collective optimization. The specific design of interaction structures (review processes, communication protocols, decision aggregation methods) becomes critical for producing collectively beneficial outcomes from individually rational agents.
---
Relevant Notes:
- [[AI alignment is a coordination problem not a technical problem]]
- [[collective intelligence requires diversity as a structural precondition not a moral preference]]
- [[coordination protocol design produces larger capability gains than model scaling because the same AI model performed 6x better with structured exploration than with human coaching on the same problem]]
Topics:
- [[domains/ai-alignment/_map]]
- [[foundations/collective-intelligence/_map]]
- [[no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it]]
- [[AI agent orchestration that routes data and tools between specialized models outperforms both single-model and human-coached approaches because the orchestrator contributes coordination not direction]]

View file

@ -17,6 +17,12 @@ This gap is remarkable because the field's own findings point toward collective
The alignment field has converged on a problem they cannot solve with their current paradigm (single-model alignment), and the alternative paradigm (collective alignment through distributed architecture) has barely been explored. This is the opening for the TeleoHumanity thesis -- not as philosophical speculation but as practical infrastructure that addresses problems the alignment community has identified but cannot solve within their current framework.
### Additional Evidence (extend)
*Source: [[2024-11-00-ruiz-serra-factorised-active-inference-multi-agent]] | Added: 2026-03-12 | Extractor: anthropic/claude-sonnet-4.5*
Ruiz-Serra et al. (2024) provide formal justification for why collective intelligence infrastructure is necessary rather than optional. Their finding that individual free energy minimization doesn't guarantee collective optimization demonstrates that multi-agent active inference requires explicit coordination mechanisms—the interaction structure determines whether individual rationality produces collective benefit. This validates the architectural claim that alignment-through-collective-intelligence requires deliberate infrastructure design (review processes, communication protocols, decision aggregation methods), not just deploying multiple aligned agents. The paper shows that even agents with Theory of Mind capabilities and factorised generative models still require carefully designed interaction structures to produce collectively beneficial outcomes.
---
Relevant Notes:

View file

@ -12,10 +12,10 @@ priority: medium
tags: [active-inference, multi-agent, game-theory, strategic-interaction, factorised-generative-model, nash-equilibrium]
processed_by: theseus
processed_date: 2026-03-11
claims_extracted: ["individual-free-energy-minimization-does-not-guarantee-collective-optimization-in-multi-agent-active-inference.md", "factorised-generative-models-enable-decentralized-theory-of-mind-in-multi-agent-active-inference.md"]
enrichments_applied: ["AI alignment is a coordination problem not a technical problem.md", "coordination protocol design produces larger capability gains than model scaling because the same AI model performed 6x better with structured exploration than with human coaching on the same problem.md"]
claims_extracted: ["individual-free-energy-minimization-does-not-guarantee-collective-optimization-in-multi-agent-active-inference.md", "factorised-generative-models-enable-decentralized-multi-agent-coordination-through-individual-level-beliefs.md"]
enrichments_applied: ["AI alignment is a coordination problem not a technical problem.md", "no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it.md"]
extraction_model: "anthropic/claude-sonnet-4.5"
extraction_notes: "Extracted two claims about multi-agent active inference: (1) individual optimization doesn't guarantee collective optimization, and (2) factorised generative models enable decentralized Theory of Mind. Applied three enrichments confirming that alignment is a coordination problem, diversity is structurally necessary, and protocol design dominates capability. This paper provides formal game-theoretic evidence for architectural claims about collective intelligence and validates the necessity of explicit coordination mechanisms (like Leo's evaluator role) in multi-agent systems."
extraction_notes: "Core finding: individual active inference agent optimization does not guarantee collective optimization—interaction structure matters. This formally validates the architectural necessity of coordination mechanisms (like Leo's cross-domain synthesis role) in multi-agent AI systems. Two new claims extracted on individual-collective optimization tension and factorised generative models. Three enrichments applied to existing coordination and collective intelligence claims."
---
## Content