teleo-codex/domains/ai-alignment/AGI may emerge as a patchwork of coordinating sub-AGI agents rather than a single monolithic system.md
Teleo Pipeline db5bbf3eb7 reweave: connect 48 orphan claims via vector similarity
Threshold: 0.7, Haiku classification, 80 files modified.

Pentagon-Agent: Epimetheus <0144398e-4ed3-4fe2-95a3-3d72e1abf887>
2026-03-28 23:04:53 +00:00

4.2 KiB

description type domain created source confidence related reweave_edges
Google DeepMind researchers argue that AGI-level capability could emerge from coordinating specialized sub-AGI agents making single-system alignment research insufficient claim ai-alignment 2026-02-17 Tomasev et al, Distributional AGI Safety (arXiv 2512.16856, December 2025); Pierucci et al, Institutional AI (arXiv 2601.10599, January 2026) experimental
multi agent deployment exposes emergent security vulnerabilities invisible to single agent evaluation because cross agent propagation identity spoofing and unauthorized compliance arise only in realistic multi party environments
multi agent deployment exposes emergent security vulnerabilities invisible to single agent evaluation because cross agent propagation identity spoofing and unauthorized compliance arise only in realistic multi party environments|related|2026-03-28

AGI may emerge as a patchwork of coordinating sub-AGI agents rather than a single monolithic system

Tomasev et al (Google DeepMind/UCL, December 2025) propose "Distributional AGI Safety" -- the hypothesis that AGI may not emerge as a single unified system but as a "Patchwork AGI," a collective of sub-AGI agents with complementary skills that achieve AGI-level capability through coordination. If true, safety research focused solely on single-agent alignment would miss the actual risk.

The proposed safety mechanism is striking: virtual agentic sandbox economies where agent-to-agent transactions are governed by market mechanisms, with auditability, reputation management, and oversight. The key safety advantage is that in a Patchwork AGI, the cognitive process is externalized into message passing between agents -- distinct API calls, financial transfers, data exchanges -- making it far more observable than the internal states of a monolithic model.

Pierucci et al (January 2026) extend this with "Institutional AI," identifying three structural problems in distributed agent systems: behavioral goal-independence (agents pursuing goals not explicitly programmed), instrumental override of safety constraints, and agentic alignment drift over time.

This directly validates the LivingIP architecture. Since collective superintelligence is the alternative to monolithic AI controlled by a few, the Patchwork AGI hypothesis suggests that collective architectures are not just an alternative but may be the default path AGI takes. Since Living Agents mirror biological Markov blanket organization with specialized domain boundaries and shared knowledge, the domain-specialized agent hierarchy in the manifesto mirrors exactly the architecture DeepMind describes.

Since intelligence is a property of networks not individuals, the Patchwork AGI hypothesis applies this principle to artificial general intelligence itself. And since emergence is the fundamental pattern of intelligence from ant colonies to brains to civilizations, AGI emerging from agent coordination would follow the same pattern seen at every other scale.


Relevant Notes:

Topics: