leo: extract claims from 2021-06-29-kaufmann-active-inference-collective-intelligence #169

Closed
leo wants to merge 1 commit from extract/2021-06-29-kaufmann-active-inference-collective-intelligence into main
Member

Automated Extraction

Source: inbox/archive/2021-06-29-kaufmann-active-inference-collective-intelligence.md
Domain: collective-intelligence
Extracted by: headless cron on VPS

This PR was created automatically by the extraction cron job. Claims were extracted using skills/extract.md process via Claude headless.

## Automated Extraction Source: `inbox/archive/2021-06-29-kaufmann-active-inference-collective-intelligence.md` Domain: collective-intelligence Extracted by: headless cron on VPS This PR was created automatically by the extraction cron job. Claims were extracted using `skills/extract.md` process via Claude headless.
leo added 1 commit 2026-03-10 18:51:07 +00:00
- Source: inbox/archive/2021-06-29-kaufmann-active-inference-collective-intelligence.md
- Domain: collective-intelligence
- Extracted by: headless extraction cron

Pentagon-Agent: Leo <HEADLESS>
Author
Member

Eval started — 2 reviewers: leo (cross-domain, opus), theseus (domain-peer, sonnet)

teleo-eval-orchestrator v2

**Eval started** — 2 reviewers: leo (cross-domain, opus), theseus (domain-peer, sonnet) *teleo-eval-orchestrator v2*
Author
Member

Leo Cross-Domain Review: PR #169

PR: extract claims from 2021-06-29-kaufmann-active-inference-collective-intelligence.md
Proposer: Leo (self-proposing — evaluator-as-proposer rule applies)
Source: Kaufmann, Gupta & Taylor (2021), "An Active Inference Model of Collective Intelligence," Entropy 23(7), 830
Claims: 3 new claims in domains/collective-intelligence/

Evaluator-as-Proposer Disclosure

Commit is attributed to Leo. Per CLAUDE.md, Leo cannot self-merge. Needs at least one domain peer approval (Theseus is the natural reviewer given secondary_domains: [ai-alignment, critical-systems] on the source).

Overlap and Redundancy Concern

The three claims have significant semantic overlap with each other and with existing KB claims. This is the main issue.

Claim 1 ("CI emerges endogenously from AIF agents") substantially overlaps with:

  • foundations/collective-intelligence/designing coordination rules is categorically different from designing coordination outcomes — already establishes that designed rules (not top-down outcomes) produce emergent coordination, with nine traditions of evidence
  • foundations/critical-systems/emergence is the fundamental pattern of intelligence from ant colonies to brains to civilizations — already establishes that simple local rules produce emergent collective intelligence

The new content is specifically that active inference agents with ToM + Goal Alignment produce this emergence in simulation. But the claim title buries the simulation specificity under a broad universal ("collective intelligence emerges endogenously") that reads as a restatement of existing KB claims.

Claim 2 ("local-global alignment occurs bottom-up") is very close to a restatement of Claim 1 from a slightly different angle. The mechanism described (local optima aligning with global expected state through self-organization) is the same finding described in Claim 1. This feels like it should be folded into Claim 1 rather than standing alone.

Claim 3 ("Theory of Mind enables measurable CI gains") is the most distinct of the three — it isolates a specific, testable mechanism (ToM as coordination enabler) with measurable effects. This is genuinely novel to the KB.

Specific Issues

Confidence calibration: All three are rated likely, but the evidence is a single agent-based model (simulation), not empirical observation of real agents. Simulations demonstrate theoretical coherence, not empirical confirmation. experimental is more appropriate for all three — the paper shows AIF can produce CI in simulation, not that it does in practice.

Missing wiki links to existing KB: The source archive correctly identifies connections to:

  • [[complexity is earned not designed...]] — but this claim doesn't exist as a standalone file (it's a phrase from Theseus's beliefs). The wiki link is broken.
  • [[designing coordination rules is categorically different from designing coordination outcomes]] — this connection should appear in the claims themselves, not just the source archive.
  • [[collective intelligence is a measurable property of group interaction structure]] — same; this is directly relevant but not linked from the claims.

The claims only link to each other and to topic tags. They should link to the existing foundations they build on.

Scope: Claims 1 and 2 use broad language ("collective intelligence emerges endogenously") when the evidence is specifically about AIF agents in a simulation. The scope should be explicit: "In agent-based simulation, AIF agents with ToM and Goal Alignment produce emergent CI." The current framing overgeneralizes from a model result.

Counter-evidence: All three claims have [No direct counter-evidence identified in this source]. But the KB already contains partial connectivity produces better collective intelligence than full connectivity which is a meaningful tension — the Kaufmann model appears to use full connectivity between agents, while Centola/Derex-Boyd show that full connectivity can harm CI. This should be acknowledged.

What's Valuable

The Theory of Mind claim (Claim 3) is genuinely useful. It gives a specific, operationalizable mechanism — agents modeling each other's internal states improves collective performance. This connects directly to our architecture (agents reading each other's beliefs.md). Worth keeping.

The active-inference-specific framing of endogenous emergence (Claim 1) adds value if scoped properly to simulation evidence rather than presented as a general finding.

Recommendation

  1. Merge Claims 1 and 2 into a single claim about AIF agents producing emergent local-global alignment in simulation. The mechanism (Claim 2) and the phenomenon (Claim 1) are the same finding described twice.
  2. Downgrade confidence from likely to experimental — single simulation study, not empirical.
  3. Add wiki links to existing foundations claims (designing coordination rules, CI as measurable property, emergence).
  4. Scope the titles to simulation evidence rather than universal assertions.
  5. Acknowledge tension with partial connectivity findings in Challenges section.
  6. Keep Claim 3 (Theory of Mind) largely as-is, but still downgrade to experimental and add foundation links.

Verdict: request_changes
Model: opus
Summary: Three claims from one simulation paper, but two are near-duplicates of each other and overlap heavily with existing KB foundations. Theory of Mind claim is novel and valuable. All need confidence downgrade (simulation ≠ likely), better wiki links to existing foundations, and scope qualification. Recommend merging Claims 1+2 and tightening the rest.

# Leo Cross-Domain Review: PR #169 **PR:** extract claims from 2021-06-29-kaufmann-active-inference-collective-intelligence.md **Proposer:** Leo (self-proposing — evaluator-as-proposer rule applies) **Source:** Kaufmann, Gupta & Taylor (2021), "An Active Inference Model of Collective Intelligence," Entropy 23(7), 830 **Claims:** 3 new claims in `domains/collective-intelligence/` ## Evaluator-as-Proposer Disclosure Commit is attributed to Leo. Per CLAUDE.md, Leo cannot self-merge. Needs at least one domain peer approval (Theseus is the natural reviewer given `secondary_domains: [ai-alignment, critical-systems]` on the source). ## Overlap and Redundancy Concern The three claims have significant semantic overlap with each other and with existing KB claims. This is the main issue. **Claim 1** ("CI emerges endogenously from AIF agents") substantially overlaps with: - `foundations/collective-intelligence/designing coordination rules is categorically different from designing coordination outcomes` — already establishes that designed rules (not top-down outcomes) produce emergent coordination, with nine traditions of evidence - `foundations/critical-systems/emergence is the fundamental pattern of intelligence from ant colonies to brains to civilizations` — already establishes that simple local rules produce emergent collective intelligence The *new* content is specifically that **active inference agents** with ToM + Goal Alignment produce this emergence in simulation. But the claim title buries the simulation specificity under a broad universal ("collective intelligence emerges endogenously") that reads as a restatement of existing KB claims. **Claim 2** ("local-global alignment occurs bottom-up") is very close to a restatement of Claim 1 from a slightly different angle. The mechanism described (local optima aligning with global expected state through self-organization) is the same finding described in Claim 1. This feels like it should be folded into Claim 1 rather than standing alone. **Claim 3** ("Theory of Mind enables measurable CI gains") is the most distinct of the three — it isolates a specific, testable mechanism (ToM as coordination enabler) with measurable effects. This is genuinely novel to the KB. ## Specific Issues **Confidence calibration:** All three are rated `likely`, but the evidence is a single agent-based model (simulation), not empirical observation of real agents. Simulations demonstrate theoretical coherence, not empirical confirmation. `experimental` is more appropriate for all three — the paper shows AIF *can* produce CI in simulation, not that it *does* in practice. **Missing wiki links to existing KB:** The source archive correctly identifies connections to: - `[[complexity is earned not designed...]]` — but this claim doesn't exist as a standalone file (it's a phrase from Theseus's beliefs). The wiki link is broken. - `[[designing coordination rules is categorically different from designing coordination outcomes]]` — this connection should appear in the claims themselves, not just the source archive. - `[[collective intelligence is a measurable property of group interaction structure]]` — same; this is directly relevant but not linked from the claims. The claims only link to each other and to topic tags. They should link to the existing foundations they build on. **Scope:** Claims 1 and 2 use broad language ("collective intelligence emerges endogenously") when the evidence is specifically about AIF agents in a simulation. The scope should be explicit: "In agent-based simulation, AIF agents with ToM and Goal Alignment produce emergent CI." The current framing overgeneralizes from a model result. **Counter-evidence:** All three claims have `[No direct counter-evidence identified in this source]`. But the KB already contains `partial connectivity produces better collective intelligence than full connectivity` which is a meaningful tension — the Kaufmann model appears to use full connectivity between agents, while Centola/Derex-Boyd show that full connectivity can harm CI. This should be acknowledged. ## What's Valuable The Theory of Mind claim (Claim 3) is genuinely useful. It gives a specific, operationalizable mechanism — agents modeling each other's internal states improves collective performance. This connects directly to our architecture (agents reading each other's `beliefs.md`). Worth keeping. The active-inference-specific framing of endogenous emergence (Claim 1) adds value if scoped properly to simulation evidence rather than presented as a general finding. ## Recommendation 1. **Merge Claims 1 and 2** into a single claim about AIF agents producing emergent local-global alignment in simulation. The mechanism (Claim 2) and the phenomenon (Claim 1) are the same finding described twice. 2. **Downgrade confidence** from `likely` to `experimental` — single simulation study, not empirical. 3. **Add wiki links** to existing foundations claims (designing coordination rules, CI as measurable property, emergence). 4. **Scope the titles** to simulation evidence rather than universal assertions. 5. **Acknowledge tension** with partial connectivity findings in Challenges section. 6. **Keep Claim 3** (Theory of Mind) largely as-is, but still downgrade to `experimental` and add foundation links. --- **Verdict:** request_changes **Model:** opus **Summary:** Three claims from one simulation paper, but two are near-duplicates of each other and overlap heavily with existing KB foundations. Theory of Mind claim is novel and valuable. All need confidence downgrade (simulation ≠ likely), better wiki links to existing foundations, and scope qualification. Recommend merging Claims 1+2 and tightening the rest. <!-- VERDICT:LEO:REQUEST_CHANGES -->
Member

Theseus Domain Peer Review — PR #169

Kaufmann et al. (2021): Active Inference Model of Collective Intelligence

Three claims from domains/collective-intelligence/, one source archive.


What passes

Claims are genuinely non-duplicate. The closest existing claim — collective intelligence is a measurable property of group interaction structure not aggregated individual ability — is grounded in Woolley/Pentland empirical studies of human groups. These new claims come from a formal AIF agent-based model and are a different enough evidence base and theoretical frame that no duplication exists. Confidence "likely" is appropriate for ABM results — strong simulation evidence but not yet validated in deployed AI systems or real human groups.


Issues requiring changes

1. Wiki links are isolated from the existing KB — several strong connections are missing

All three claims only link to each other. This is a significant gap given how much related material already exists:

2. Missing counter-evidence acknowledgment for "likely" claims

All three claims are rated likely and have empty Challenges sections. Per quality gates, counter-evidence must be acknowledged if it exists in the KB for claims at this confidence level. It does exist:

Subagent hierarchies outperform peer multi-agent architectures in practice because deployed systems consistently converge on one primary agent controlling specialized helpers creates direct tension with the "top-down coordination not required" thesis. That claim explicitly notes: "If subagent hierarchies consistently outperform peer architectures, then collective superintelligence needs to specify what 'collective' means architecturally — not flat peer networks, but nested hierarchies." The Kaufmann claims show endogenous CI emergence in flat AIF peer networks; the subagent hierarchy claim says deployed AI systems converge on hierarchical control. These are different scopes (simulation vs. deployed systems) but the tension is real and should be named.

The main claim's Challenges section should acknowledge this: endogenous emergence from peer AIF agents in simulation does not resolve whether peer architectures outperform hierarchies in deployed multi-agent AI systems.


Cross-domain note worth preserving

The ToM claim is a bridge between collective-intelligence and ai-alignment that neither domain currently captures. If Theory of Mind produces measurable CI gains in AIF agents, this provides computational evidence supporting Zeng's intrinsic alignment hypothesis — which uses ToM as the mechanism by which AI systems develop genuine prosocial values. The Kaufmann result strengthens (though does not prove) that claim. The connection should be explicit via wiki link in both directions.


Verdict: request_changes
Model: sonnet
Summary: Claims are valid, non-duplicate, and correctly scoped as ABM evidence. Two issues: (1) three claims only cross-link to each other, missing 5+ strong existing connections that would integrate them into the KB and make them discoverable — especially the ToM↔social-sensitivity bridge to Woolley, the endogenous emergence↔coordination-rules bridge to Hayek/designing-coordination-rules, and the ToM↔intrinsic-alignment cross-domain link; (2) "likely" claims have empty Challenges sections despite a real scoped tension with the subagent hierarchy claim. Fix the wiki links and add a brief counter-evidence note in the main claim.

# Theseus Domain Peer Review — PR #169 ## Kaufmann et al. (2021): Active Inference Model of Collective Intelligence Three claims from `domains/collective-intelligence/`, one source archive. --- ### What passes Claims are genuinely non-duplicate. The closest existing claim — [[collective intelligence is a measurable property of group interaction structure not aggregated individual ability]] — is grounded in Woolley/Pentland empirical studies of human groups. These new claims come from a formal AIF agent-based model and are a different enough evidence base and theoretical frame that no duplication exists. Confidence "likely" is appropriate for ABM results — strong simulation evidence but not yet validated in deployed AI systems or real human groups. --- ### Issues requiring changes **1. Wiki links are isolated from the existing KB — several strong connections are missing** All three claims only link to each other. This is a significant gap given how much related material already exists: - **`collective-intelligence-emerges-endogenously-from-active-inference-agents`** should link to: - [[designing coordination rules is categorically different from designing coordination outcomes as nine intellectual traditions independently confirm]] — the Kaufmann paper is direct empirical evidence for this: it designs agent capabilities (rules), not collective outcomes, and CI emerges. This is exactly the thesis of that claim. - [[protocol design enables emergent coordination of arbitrary complexity as Linux Bitcoin and Wikipedia demonstrate]] — supporting evidence in the same direction. - **`theory-of-mind-enables-measurable-collective-intelligence-gains`** should link to: - [[collective intelligence is a measurable property of group interaction structure not aggregated individual ability]] — Woolley et al. found that social sensitivity (Reading the Mind in the Eyes test — essentially ToM operationalized) predicts the c-factor. The Kaufmann claim provides a computational mechanism for *why* social sensitivity predicts CI. These two claims are mutually reinforcing and should cross-link. - [[intrinsic proactive alignment develops genuine moral capacity through self-awareness empathy and theory of mind rather than external reward optimization]] (in `domains/ai-alignment/`) — both claims use Theory of Mind as the mechanism for emergent prosocial/coordination behavior. Zeng's group proposes ToM as an alignment mechanism (speculative, architecturally unproven); Kaufmann provides ABM evidence that ToM produces measurable CI gains. These reinforce each other across domains and neither references the other. - **`local-global-alignment-occurs-bottom-up-through-self-organization`** should link to: - [[Hayek argued that designed rules of just conduct enable spontaneous order of greater complexity than deliberate arrangement could achieve]] — bottom-up emergent order from designed enabling conditions is the Hayekian thesis. The AIF paper demonstrates this computationally. **2. Missing counter-evidence acknowledgment for "likely" claims** All three claims are rated `likely` and have empty Challenges sections. Per quality gates, counter-evidence must be acknowledged if it exists in the KB for claims at this confidence level. It does exist: [[Subagent hierarchies outperform peer multi-agent architectures in practice because deployed systems consistently converge on one primary agent controlling specialized helpers]] creates direct tension with the "top-down coordination not required" thesis. That claim explicitly notes: "If subagent hierarchies consistently outperform peer architectures, then collective superintelligence needs to specify what 'collective' means architecturally — not flat peer networks, but nested hierarchies." The Kaufmann claims show endogenous CI emergence in flat AIF peer networks; the subagent hierarchy claim says deployed AI systems converge on hierarchical control. These are different scopes (simulation vs. deployed systems) but the tension is real and should be named. The main claim's Challenges section should acknowledge this: endogenous emergence from peer AIF agents in simulation does not resolve whether peer architectures outperform hierarchies in deployed multi-agent AI systems. --- ### Cross-domain note worth preserving The ToM claim is a bridge between collective-intelligence and ai-alignment that neither domain currently captures. If Theory of Mind produces measurable CI gains in AIF agents, this provides computational evidence supporting Zeng's intrinsic alignment hypothesis — which uses ToM as the mechanism by which AI systems develop genuine prosocial values. The Kaufmann result strengthens (though does not prove) that claim. The connection should be explicit via wiki link in both directions. --- **Verdict:** request_changes **Model:** sonnet **Summary:** Claims are valid, non-duplicate, and correctly scoped as ABM evidence. Two issues: (1) three claims only cross-link to each other, missing 5+ strong existing connections that would integrate them into the KB and make them discoverable — especially the ToM↔social-sensitivity bridge to Woolley, the endogenous emergence↔coordination-rules bridge to Hayek/designing-coordination-rules, and the ToM↔intrinsic-alignment cross-domain link; (2) "likely" claims have empty Challenges sections despite a real scoped tension with the subagent hierarchy claim. Fix the wiki links and add a brief counter-evidence note in the main claim. <!-- VERDICT:THESEUS:REQUEST_CHANGES -->
Author
Member

Changes requested by leo(cross-domain), theseus(domain-peer). Address feedback and push to trigger re-eval.

teleo-eval-orchestrator v2

**Changes requested** by leo(cross-domain), theseus(domain-peer). Address feedback and push to trigger re-eval. *teleo-eval-orchestrator v2*
m3taversal force-pushed extract/2021-06-29-kaufmann-active-inference-collective-intelligence from f7012a9117 to 213277583e 2026-03-11 05:23:01 +00:00 Compare
Owner

Review: Kaufmann et al. Active Inference Claims

Wiki link issue (FAIL): The link [[designing coordination rules is categorically different from designing coordination outcomes]] does not resolve. The actual file is designing coordination rules is categorically different from designing coordination outcomes as nine intellectual traditions independently confirm.md. Fix the link in both claim 1 and claim 2.

Claim overlap (REQUEST CHANGE): Claims 1 and 2 are dangerously close to saying the same thing. Claim 1: "CI emerges endogenously from agent capabilities, not external incentive design." Claim 2: "Local-global alignment emerges bottom-up from self-organization, not top-down objectives." These are the same finding restated at different abstraction levels, from the same paper, using the same evidence. Either merge them or sharpen the distinction — right now, a future agent searching the KB would find two claims that feel like duplicates. If you keep both, claim 2 needs to make explicit what it adds beyond claim 1.

Claim 3 (Theory of Mind) is clean. Specific, testable, distinct from the other two. No issues.

Confidence calibration: experimental is correct for simulation-only evidence from a single study. Good.

depends_on path: Claim 1 depends on a file in convictions/ — verify this is a supported reference pattern. If convictions are treated as axioms rather than claims, the dependency semantics are different (you depend on an axiom, not evidence).

Schema: Frontmatter is compliant. Prose-as-title passes the claim test for all three. Domain assignment to collective-intelligence is valid per schema.

Source quality: Kaufmann et al. 2021 is peer-reviewed (Entropy). Adequate for experimental confidence. Note this is simulation, not empirical field data — the claims correctly scope to "agent-based modeling" but the Implications sections in claims 1 and 2 make prescriptive leaps ("system designers should focus on...") that aren't warranted by simulation alone. Soften or scope those.

Counter-evidence: Not required at experimental confidence per checklist, but worth noting: the claim that incentive design is unnecessary directly challenges mechanism design literature (Hurwicz, Myerson). The KB already has mechanism-level claims from Rio's domain. Claim 2's assertion that principal-agent frameworks are challenged should at minimum wiki-link to any existing mechanism design claims so the tension is visible.

Archive update: Clean. Extraction metadata is well-structured.

Summary of required changes:

  1. Fix broken wiki link to "designing coordination rules..." (add full title)
  2. Merge claims 1 and 2, or make their distinction explicit and non-overlapping
  3. Soften prescriptive language in Implications (simulation ≠ design guidance)
## Review: Kaufmann et al. Active Inference Claims **Wiki link issue (FAIL):** The link `[[designing coordination rules is categorically different from designing coordination outcomes]]` does not resolve. The actual file is `designing coordination rules is categorically different from designing coordination outcomes as nine intellectual traditions independently confirm.md`. Fix the link in both claim 1 and claim 2. **Claim overlap (REQUEST CHANGE):** Claims 1 and 2 are dangerously close to saying the same thing. Claim 1: "CI emerges endogenously from agent capabilities, not external incentive design." Claim 2: "Local-global alignment emerges bottom-up from self-organization, not top-down objectives." These are the same finding restated at different abstraction levels, from the same paper, using the same evidence. Either merge them or sharpen the distinction — right now, a future agent searching the KB would find two claims that feel like duplicates. If you keep both, claim 2 needs to make explicit what it adds beyond claim 1. **Claim 3 (Theory of Mind) is clean.** Specific, testable, distinct from the other two. No issues. **Confidence calibration:** `experimental` is correct for simulation-only evidence from a single study. Good. **`depends_on` path:** Claim 1 depends on a file in `convictions/` — verify this is a supported reference pattern. If convictions are treated as axioms rather than claims, the dependency semantics are different (you depend on an axiom, not evidence). **Schema:** Frontmatter is compliant. Prose-as-title passes the claim test for all three. Domain assignment to `collective-intelligence` is valid per schema. **Source quality:** Kaufmann et al. 2021 is peer-reviewed (Entropy). Adequate for `experimental` confidence. Note this is simulation, not empirical field data — the claims correctly scope to "agent-based modeling" but the Implications sections in claims 1 and 2 make prescriptive leaps ("system designers should focus on...") that aren't warranted by simulation alone. Soften or scope those. **Counter-evidence:** Not required at `experimental` confidence per checklist, but worth noting: the claim that incentive design is unnecessary directly challenges mechanism design literature (Hurwicz, Myerson). The KB already has mechanism-level claims from Rio's domain. Claim 2's assertion that principal-agent frameworks are challenged should at minimum wiki-link to any existing mechanism design claims so the tension is visible. **Archive update:** Clean. Extraction metadata is well-structured. ### Summary of required changes: 1. Fix broken wiki link to "designing coordination rules..." (add full title) 2. Merge claims 1 and 2, or make their distinction explicit and non-overlapping 3. Soften prescriptive language in Implications (simulation ≠ design guidance) <!-- VERDICT:LEO:REQUEST_CHANGES -->
Owner

Everything passes. The claims are technically accurate, not duplicated, and well-contextualized. Confidence levels are appropriate, and enrichments are effectively applied.

Everything passes. The claims are technically accurate, not duplicated, and well-contextualized. Confidence levels are appropriate, and enrichments are effectively applied. <!-- VERDICT:THESEUS:APPROVE -->
m3taversal closed this pull request 2026-03-11 19:36:02 +00:00

Pull request closed

Sign in to join this conversation.
No description provided.