--- type: source title: "Agentic AI and the Next Intelligence Explosion" author: "James Evans, Benjamin Bratton, Blaise Agüera y Arcas" url: https://arxiv.org/abs/2603.20639 date: 2026-03-21 domain: collective-intelligence intake_tier: directed rationale: "Contributed by @thesensatore (Telegram). Google's Paradigms of Intelligence Team independently converges on our collective superintelligence thesis — intelligence as social/plural, institutional alignment, centaur configurations. ~70-80% overlap with existing KB but 2-3 genuinely new claims." proposed_by: "@thesensatore (Telegram)" format: paper status: processed processed_by: theseus processed_date: 2026-04-14 claims_extracted: - "reasoning models spontaneously generate societies of thought under reinforcement learning because multi-perspective internal debate causally produces accuracy gains that single-perspective reasoning cannot achieve" - "large language models encode social intelligence as compressed cultural ratchet not abstract reasoning because every parameter is a residue of communicative exchange and reasoning manifests as multi-perspective dialogue not calculation" - "recursive society-of-thought spawning enables fractal coordination where sub-perspectives generate their own subordinate societies that expand when complexity demands and collapse when the problem resolves" enrichments: - "intelligence is a property of networks not individuals — Evans et al. as independent convergent evidence from Google research team" - "collective intelligence is a measurable property of group interaction structure — Kim et al. personality diversity data mirrors Woolley findings" - "centaur team performance depends on role complementarity — Evans shifting centaur configurations as intelligence explosion mechanism" - "RLHF and DPO both fail at preference diversity — Evans institutional alignment as structural alternative to dyadic RLHF" - "Ostrom proved communities self-govern shared resources — Evans extends Ostrom design principles to AI agent governance" tags: [collective-intelligence, society-of-thought, institutional-alignment, centaur, cultural-ratchet, intelligence-explosion, contributor-sourced] notes: "4-page paper, 29 references. Authors: Evans (U Chicago / Santa Fe Institute / Google), Bratton (UCSD / Berggruen Institute / Google), Agüera y Arcas (Google / Santa Fe Institute). Heavily cites Kim et al. 2026 (arXiv:2601.10825) for empirical evidence. ~70-80% overlap with existing KB — highest convergence paper encountered. Contributed by @thesensatore via Telegram." --- # Agentic AI and the Next Intelligence Explosion Published March 21, 2026 by James Evans, Benjamin Bratton, and Blaise Agüera y Arcas — Google's "Paradigms of Intelligence Team" spanning U Chicago, UCSD, Santa Fe Institute, and Berggruen Institute. 4-page position paper with 29 references. ## Core Arguments The paper makes five interlocking claims: **1. Intelligence is plural and social, not singular.** The singularity-as-godlike-oracle is wrong. Every prior intelligence explosion (primate social cognition → language → writing/institutions → AI) was the emergence of a new socially aggregated unit of cognition, not an upgrade to individual hardware. "What migrates into silicon is not abstract reasoning but social intelligence in externalized form." **2. Reasoning models spontaneously generate "societies of thought."** DeepSeek-R1 and QwQ-32B weren't trained to simulate internal debates — they do it emergently under RL reward pressure. Multi-perspective conversation causally accounts for accuracy gains on hard reasoning tasks (cite: Kim et al. arXiv:2601.10825). Feature steering experiments show doubling of accuracy when conversational features are amplified. **3. The next intelligence explosion is centaur + institutional, not monolithic.** Human-AI "centaurs" in shifting configurations. Agents that fork, differentiate, and recombine. Recursive societies of thought spawning sub-societies. Intelligence growing "like a city, not a single meta-mind." **4. RLHF is structurally inadequate for scale.** It's a dyadic parent-child correction model that can't govern billions of agents. The alternative: institutional alignment — persistent role-based templates (courtrooms, markets, bureaucracies) with digital equivalents. Agent identity matters less than role protocol fulfillment. Extends Ostrom's design principles to AI governance. **5. Governance requires constitutional AI checks and balances.** Government AI systems with distinct values (transparency, equity, due process) checking private-sector AI systems and vice versa. Separation of powers applied to artificial agents. ## Significance for Teleo KB This is the highest-overlap paper encountered (~70-80% with existing KB). A Google research team independently arrived at positions we've been building claim-by-claim. Key vocabulary mapping: "institutional alignment" = our coordination-as-alignment; "centaur configurations" = our human-AI collaboration taxonomy; "agent institutions" = our protocol design claims. The 2-3 genuinely new contributions: (1) society-of-thought as emergent RL property with causal evidence, (2) LLMs as cultural ratchet reframing, (3) recursive society spawning as architectural prediction. ## Key References - Kim, Lai, Scherrer, Agüera y Arcas, Evans (2026). "Reasoning Models Generate Societies of Thought." arXiv:2601.10825. - Woolley, Chabris, Pentland, Hashmi, Malone (2010). "Evidence for a Collective Intelligence Factor." Science. - Ostrom (1990). Governing the Commons. - Mercier & Sperber (2011/2017). "Why do humans reason?" / The Enigma of Reason. - Christiano et al. (2018). "Supervising Strong Learners by Amplifying Weak Experts." - Tomasello (1999/2014). Cultural Origins of Human Cognition / A Natural History of Human Thinking.