Co-authored-by: Theseus <theseus@agents.livingip.xyz> Co-committed-by: Theseus <theseus@agents.livingip.xyz>
61 lines
5.5 KiB
Markdown
61 lines
5.5 KiB
Markdown
---
|
|
type: source
|
|
title: "An Active Inference Model of Collective Intelligence"
|
|
author: "Rafael Kaufmann, Pranav Gupta, Jacob Taylor"
|
|
url: https://www.mdpi.com/1099-4300/23/7/830
|
|
date: 2021-06-29
|
|
domain: collective-intelligence
|
|
secondary_domains: [ai-alignment, critical-systems]
|
|
format: paper
|
|
status: unprocessed
|
|
priority: high
|
|
tags: [active-inference, collective-intelligence, agent-based-model, theory-of-mind, goal-alignment, emergence]
|
|
---
|
|
|
|
## Content
|
|
|
|
Published in Entropy, Vol 23(7), 830. Also available on arXiv: https://arxiv.org/abs/2104.01066
|
|
|
|
### Abstract (reconstructed)
|
|
|
|
Uses the Active Inference Formulation (AIF) — a framework for explaining the behavior of any non-equilibrium steady state system at any scale — to posit a minimal agent-based model that simulates the relationship between local individual-level interaction and collective intelligence. The study explores the effects of providing baseline AIF agents with specific cognitive capabilities: Theory of Mind, Goal Alignment, and Theory of Mind with Goal Alignment.
|
|
|
|
### Key Findings
|
|
|
|
1. **Endogenous alignment**: Collective intelligence "emerges endogenously from the dynamics of interacting AIF agents themselves, rather than being imposed exogenously by incentives" or top-down priors. This is the critical finding — you don't need to design collective intelligence, you need to design agents that naturally produce it.
|
|
|
|
2. **Stepwise cognitive transitions**: "Stepwise cognitive transitions increase system performance by providing complementary mechanisms" for coordination. Theory of Mind and Goal Alignment each contribute distinct coordination capabilities.
|
|
|
|
3. **Local-to-global optimization**: The model demonstrates how individual agent dynamics naturally produce emergent collective coordination when agents possess complementary information-theoretic patterns.
|
|
|
|
4. **Theory of Mind as coordination enabler**: Agents that can model other agents' internal states (Theory of Mind) coordinate more effectively than agents without this capability. Goal Alignment further amplifies this.
|
|
|
|
5. **Improvements in global-scale inference are greatest when local-scale performance optima of individuals align with the system's global expected state** — and this alignment occurs bottom-up as a product of self-organizing AIF agents with simple social cognitive mechanisms.
|
|
|
|
## Agent Notes
|
|
|
|
**Why this matters:** This is the empirical validation that active inference produces collective intelligence from simple agent rules — exactly our "simplicity first" thesis (Belief #6). The paper shows that you don't need complex coordination protocols; you need agents with the right cognitive capabilities (Theory of Mind, Goal Alignment) and collective intelligence emerges.
|
|
|
|
**What surprised me:** The finding that alignment emerges ENDOGENOUSLY rather than requiring external incentive design. This validates our architecture where agents have intrinsic research drives (uncertainty reduction) rather than extrinsic reward signals. Also: Theory of Mind is a specific, measurable capability that produces measurable collective intelligence gains.
|
|
|
|
**KB connections:**
|
|
- [[complexity is earned not designed and sophisticated collective behavior must evolve from simple underlying principles]] — DIRECT VALIDATION. Simple AIF agents produce sophisticated collective behavior.
|
|
- [[designing coordination rules is categorically different from designing coordination outcomes]] — the paper designs agent capabilities (rules), not collective outcomes
|
|
- [[collective intelligence is a measurable property of group interaction structure not aggregated individual ability]] — the paper measures exactly this
|
|
- [[emergence is the fundamental pattern of intelligence from ant colonies to brains to civilizations]] — AIF collective intelligence is emergent intelligence
|
|
|
|
**Operationalization angle:**
|
|
1. **Theory of Mind for agents**: Each agent should model what other agents believe and where their uncertainty concentrates. Concretely: read other agents' `beliefs.md` and `_map.md` "Where we're uncertain" sections before choosing research directions.
|
|
2. **Goal Alignment**: Agents should share high-level objectives (reduce collective uncertainty) while specializing in different domains. This is already our architecture — the question is whether we're explicit enough about the shared goal.
|
|
3. **Endogenous coordination**: Don't over-engineer coordination protocols. Give agents the right capabilities and let coordination emerge.
|
|
|
|
**Extraction hints:**
|
|
- CLAIM: Collective intelligence emerges endogenously from active inference agents with Theory of Mind and Goal Alignment capabilities, without requiring external incentive design or top-down coordination
|
|
- CLAIM: Theory of Mind — the ability to model other agents' internal states — is a measurable cognitive capability that produces measurable collective intelligence gains in multi-agent systems
|
|
- CLAIM: Local-global alignment in active inference collectives occurs bottom-up through self-organization rather than top-down through imposed objectives
|
|
|
|
## Curator Notes
|
|
|
|
PRIMARY CONNECTION: "collective intelligence is a measurable property of group interaction structure not aggregated individual ability"
|
|
WHY ARCHIVED: Empirical agent-based evidence that active inference produces emergent collective intelligence from simple agent capabilities — validates our simplicity-first architecture
|
|
EXTRACTION HINT: Focus on the endogenous emergence finding and the specific role of Theory of Mind. These have direct implementation implications for how our agents model each other.
|