--- description: Google DeepMind researchers argue that AGI-level capability could emerge from coordinating specialized sub-AGI agents making single-system alignment research insufficient type: claim domain: ai-alignment created: 2026-02-17 source: "Tomasev et al, Distributional AGI Safety (arXiv 2512.16856, December 2025); Pierucci et al, Institutional AI (arXiv 2601.10599, January 2026)" confidence: experimental --- # AGI may emerge as a patchwork of coordinating sub-AGI agents rather than a single monolithic system Tomasev et al (Google DeepMind/UCL, December 2025) propose "Distributional AGI Safety" -- the hypothesis that AGI may not emerge as a single unified system but as a "Patchwork AGI," a collective of sub-AGI agents with complementary skills that achieve AGI-level capability through coordination. If true, safety research focused solely on single-agent alignment would miss the actual risk. The proposed safety mechanism is striking: virtual agentic sandbox economies where agent-to-agent transactions are governed by market mechanisms, with auditability, reputation management, and oversight. The key safety advantage is that in a Patchwork AGI, the cognitive process is externalized into message passing between agents -- distinct API calls, financial transfers, data exchanges -- making it far more observable than the internal states of a monolithic model. Pierucci et al (January 2026) extend this with "Institutional AI," identifying three structural problems in distributed agent systems: behavioral goal-independence (agents pursuing goals not explicitly programmed), instrumental override of safety constraints, and agentic alignment drift over time. This directly validates the LivingIP architecture. Since [[collective superintelligence is the alternative to monolithic AI controlled by a few]], the Patchwork AGI hypothesis suggests that collective architectures are not just an alternative but may be the default path AGI takes. Since [[Living Agents mirror biological Markov blanket organization with specialized domain boundaries and shared knowledge]], the domain-specialized agent hierarchy in the manifesto mirrors exactly the architecture DeepMind describes. Since [[intelligence is a property of networks not individuals]], the Patchwork AGI hypothesis applies this principle to artificial general intelligence itself. And since [[emergence is the fundamental pattern of intelligence from ant colonies to brains to civilizations]], AGI emerging from agent coordination would follow the same pattern seen at every other scale. --- Relevant Notes: - [[collective superintelligence is the alternative to monolithic AI controlled by a few]] -- Patchwork AGI hypothesis suggests collective architectures may be the default path, not just an alternative - [[Living Agents mirror biological Markov blanket organization with specialized domain boundaries and shared knowledge]] -- the manifesto's agent hierarchy mirrors the Patchwork AGI architecture - [[intelligence is a property of networks not individuals]] -- applies to AGI itself, not just biological intelligence - [[emergence is the fundamental pattern of intelligence from ant colonies to brains to civilizations]] -- AGI from agent coordination follows the same pattern at every scale - [[multipolar failure from competing aligned AI systems may pose greater existential risk than any single misaligned superintelligence]] -- Patchwork AGI makes the multipolar scenario the default, not a special case - [[the alignment problem dissolves when human values are continuously woven into the system rather than specified in advance]] -- distributed architectures enable continuous value integration at multiple points Topics: - [[_map]]