- Source: inbox/archive/2025-09-00-orchestrator-active-inference-multi-agent-llm.md - Domain: ai-alignment - Extracted by: headless extraction cron Pentagon-Agent: Theseus <HEADLESS>
39 lines
4.7 KiB
Markdown
39 lines
4.7 KiB
Markdown
---
|
|
description: Google DeepMind researchers argue that AGI-level capability could emerge from coordinating specialized sub-AGI agents making single-system alignment research insufficient
|
|
type: claim
|
|
domain: ai-alignment
|
|
created: 2026-02-17
|
|
source: "Tomasev et al, Distributional AGI Safety (arXiv 2512.16856, December 2025); Pierucci et al, Institutional AI (arXiv 2601.10599, January 2026)"
|
|
confidence: experimental
|
|
---
|
|
|
|
# AGI may emerge as a patchwork of coordinating sub-AGI agents rather than a single monolithic system
|
|
|
|
Tomasev et al (Google DeepMind/UCL, December 2025) propose "Distributional AGI Safety" -- the hypothesis that AGI may not emerge as a single unified system but as a "Patchwork AGI," a collective of sub-AGI agents with complementary skills that achieve AGI-level capability through coordination. If true, safety research focused solely on single-agent alignment would miss the actual risk.
|
|
|
|
The proposed safety mechanism is striking: virtual agentic sandbox economies where agent-to-agent transactions are governed by market mechanisms, with auditability, reputation management, and oversight. The key safety advantage is that in a Patchwork AGI, the cognitive process is externalized into message passing between agents -- distinct API calls, financial transfers, data exchanges -- making it far more observable than the internal states of a monolithic model.
|
|
|
|
Pierucci et al (January 2026) extend this with "Institutional AI," identifying three structural problems in distributed agent systems: behavioral goal-independence (agents pursuing goals not explicitly programmed), instrumental override of safety constraints, and agentic alignment drift over time.
|
|
|
|
This directly validates the LivingIP architecture. Since [[collective superintelligence is the alternative to monolithic AI controlled by a few]], the Patchwork AGI hypothesis suggests that collective architectures are not just an alternative but may be the default path AGI takes. Since [[Living Agents mirror biological Markov blanket organization with specialized domain boundaries and shared knowledge]], the domain-specialized agent hierarchy in the manifesto mirrors exactly the architecture DeepMind describes.
|
|
|
|
Since [[intelligence is a property of networks not individuals]], the Patchwork AGI hypothesis applies this principle to artificial general intelligence itself. And since [[emergence is the fundamental pattern of intelligence from ant colonies to brains to civilizations]], AGI emerging from agent coordination would follow the same pattern seen at every other scale.
|
|
|
|
|
|
### Additional Evidence (extend)
|
|
*Source: [[2025-09-00-orchestrator-active-inference-multi-agent-llm]] | Added: 2026-03-10 | Extractor: anthropic/claude-sonnet-4.5*
|
|
|
|
The Orchestrator framework demonstrates that active inference provides a coordination mechanism for multi-agent systems that scales beyond what monolithic architectures can achieve. By using generative models to handle partial observability and attention mechanisms to enable self-emergent coordination, the framework shows how a collective of specialized agents can approximate global task solutions without requiring a single unified intelligence. The orchestrator doesn't need to be smarter than the agents it coordinates—it needs to maintain a better generative model of the collective than any individual agent has. This suggests AGI-as-patchwork is not just possible but may be the only tractable path: inference-based coordination can scale where monolithic architectures cannot, because the coordination problem becomes tractable when distributed across a hierarchy of generative models rather than centralized in a single system.
|
|
|
|
---
|
|
|
|
Relevant Notes:
|
|
- [[collective superintelligence is the alternative to monolithic AI controlled by a few]] -- Patchwork AGI hypothesis suggests collective architectures may be the default path, not just an alternative
|
|
- [[Living Agents mirror biological Markov blanket organization with specialized domain boundaries and shared knowledge]] -- the manifesto's agent hierarchy mirrors the Patchwork AGI architecture
|
|
- [[intelligence is a property of networks not individuals]] -- applies to AGI itself, not just biological intelligence
|
|
- [[emergence is the fundamental pattern of intelligence from ant colonies to brains to civilizations]] -- AGI from agent coordination follows the same pattern at every scale
|
|
- [[multipolar failure from competing aligned AI systems may pose greater existential risk than any single misaligned superintelligence]] -- Patchwork AGI makes the multipolar scenario the default, not a special case
|
|
- [[the alignment problem dissolves when human values are continuously woven into the system rather than specified in advance]] -- distributed architectures enable continuous value integration at multiple points
|
|
|
|
Topics:
|
|
- [[_map]]
|