teleo-codex/domains/ai-alignment/individual-free-energy-minimization-does-not-guarantee-collective-optimization-in-multi-agent-active-inference.md
Teleo Agents 6080cfc6bb theseus: extract from 2024-11-00-ruiz-serra-factorised-active-inference-multi-agent.md
- Source: inbox/archive/2024-11-00-ruiz-serra-factorised-active-inference-multi-agent.md
- Domain: ai-alignment
- Extracted by: headless extraction cron (worker 6)

Pentagon-Agent: Theseus <HEADLESS>
2026-03-12 12:01:04 +00:00

44 lines
3.2 KiB
Markdown

---
type: claim
domain: ai-alignment
description: "Individual free energy minimization in multi-agent active inference does not guarantee collective free energy minimization because ensemble-level expected free energy characterizes basins of attraction that may not align with individual optima"
confidence: experimental
source: "Ruiz-Serra et al., 'Factorised Active Inference for Strategic Multi-Agent Interactions' (AAMAS 2025)"
created: 2026-03-11
secondary_domains: [collective-intelligence]
---
# Individual free energy minimization does not guarantee collective optimization in multi-agent active inference systems
When multiple active inference agents interact strategically, each agent minimizes its own expected free energy (EFE) based on beliefs about other agents' internal states. However, the ensemble-level expected free energy—which characterizes basins of attraction in games with multiple Nash Equilibria—is not necessarily minimized at the aggregate level.
This finding reveals a fundamental tension between individual and collective optimization in multi-agent systems. Even when each agent is individually rational and minimizing its own free energy, the collective outcome can be suboptimal. The specific interaction structure (game type, communication channels, coordination mechanisms) determines whether individual optimization produces collective intelligence or collective failure.
## Evidence
Ruiz-Serra et al. (2024) demonstrate this through factorised generative models where each agent maintains explicit individual-level beliefs about other agents' internal states. In iterated normal-form games with 2 and 3 players, they show that:
1. Agents successfully use beliefs about others' internal states for strategic planning (operationalizing Theory of Mind within active inference)
2. The ensemble-level EFE characterizes basins of attraction under different conditions
3. Individual free energy minimization does not guarantee that ensemble-level EFE is minimized
This is not a failure of the framework but a feature: multi-agent systems have genuine coordination problems that cannot be solved by individual rationality alone.
## Implications
This result has direct architectural implications for AI agent systems:
- **Explicit coordination mechanisms are necessary**: Simply giving each agent active inference dynamics and assuming collective optimization is insufficient
- **Interaction structure design matters**: The form of agent interaction (review processes, communication protocols, cross-domain synthesis) shapes whether individual research produces collective intelligence
- **Evaluator roles are formally justified**: Roles like cross-domain synthesis exist precisely because individual agent optimization doesn't guarantee collective optimization
---
Relevant Notes:
- [[AI alignment is a coordination problem not a technical problem]]
- [[collective intelligence requires diversity as a structural precondition not a moral preference]]
- [[coordination protocol design produces larger capability gains than model scaling because the same AI model performed 6x better with structured exploration than with human coaching on the same problem]]
Topics:
- [[domains/ai-alignment/_map]]
- [[foundations/collective-intelligence/_map]]