- Source: inbox/archive/2024-11-00-ruiz-serra-factorised-active-inference-multi-agent.md - Domain: ai-alignment - Extracted by: headless extraction cron (worker 6) Pentagon-Agent: Theseus <HEADLESS>
43 lines
3.2 KiB
Markdown
43 lines
3.2 KiB
Markdown
---
|
|
type: claim
|
|
domain: ai-alignment
|
|
description: "Agents maintain explicit individual-level beliefs about other agents' internal states through model factorisation enabling strategic planning in joint contexts"
|
|
confidence: experimental
|
|
source: "Ruiz-Serra et al., 'Factorised Active Inference for Strategic Multi-Agent Interactions' (AAMAS 2025)"
|
|
created: 2026-03-11
|
|
secondary_domains: [collective-intelligence]
|
|
---
|
|
|
|
# Factorised generative models enable Theory of Mind in active inference agents by maintaining explicit individual-level beliefs about other agents' internal states
|
|
|
|
Ruiz-Serra et al. introduce a factorisation approach where each active inference agent maintains "explicit, individual-level beliefs about the internal states of other agents" through a factorised generative model. This enables decentralized representation of the multi-agent system where each agent can model and predict the behavior of others.
|
|
|
|
This factorisation operationalizes Theory of Mind within the active inference framework: agents don't just react to observed actions but maintain beliefs about the hidden states, preferences, and likely future actions of other agents. These beliefs are used for "strategic planning in a joint context"—agents can anticipate how others will respond to their actions and plan accordingly.
|
|
|
|
The approach enables agents to navigate strategic interactions in iterated games without requiring centralized coordination or complete information sharing. Each agent's factorised model serves as a local representation of the multi-agent system sufficient for strategic decision-making.
|
|
|
|
## Evidence
|
|
|
|
- Ruiz-Serra et al. demonstrate factorised generative models in 2-player and 3-player iterated normal-form games
|
|
- The factorisation enables "decentralized representation of the multi-agent system" where each agent maintains separate beliefs about each other agent
|
|
- Agents use these individual-level beliefs for strategic planning, successfully navigating both cooperative and non-cooperative game structures
|
|
- The framework shows agents can anticipate other agents' responses and plan strategically without centralized coordination
|
|
|
|
## Relationship to Multi-Agent Architecture
|
|
|
|
This finding validates architectural choices in multi-agent systems:
|
|
|
|
1. **Agents need models of each other**: Effective coordination requires agents to maintain beliefs about other agents' states, not just observe their outputs
|
|
2. **Decentralized representation scales**: Factorised models avoid the combinatorial explosion of centralized multi-agent state spaces
|
|
3. **Strategic planning requires Theory of Mind**: Anticipating others' responses is fundamental to effective multi-agent coordination
|
|
|
|
---
|
|
|
|
Relevant Notes:
|
|
- [[AI agent orchestration that routes data and tools between specialized models outperforms both single-model and human-coached approaches because the orchestrator contributes coordination not direction]]
|
|
- [[subagent hierarchies outperform peer multi-agent architectures in practice because deployed systems consistently converge on one primary agent controlling specialized helpers]]
|
|
- [[intelligence is a property of networks not individuals]]
|
|
|
|
Topics:
|
|
- [[domains/ai-alignment/_map]]
|
|
- [[foundations/collective-intelligence/_map]]
|