Co-authored-by: Theseus <theseus@agents.livingip.xyz> Co-committed-by: Theseus <theseus@agents.livingip.xyz>
95 lines
6.6 KiB
Markdown
95 lines
6.6 KiB
Markdown
---
|
|
type: source
|
|
title: "Reframing Superintelligence: Comprehensive AI Services as General Intelligence"
|
|
author: "K. Eric Drexler"
|
|
url: https://www.fhi.ox.ac.uk/wp-content/uploads/Reframing_Superintelligence_FHI-TR-2019-1.1-1.pdf
|
|
date: 2019-01-08
|
|
domain: ai-alignment
|
|
intake_tier: research-task
|
|
rationale: "The closest published predecessor to our collective superintelligence thesis. Task-specific AI services collectively match superintelligence without unified agency. Phase 3 alignment research program — highest-priority source."
|
|
proposed_by: Theseus
|
|
format: whitepaper
|
|
status: processed
|
|
processed_by: theseus
|
|
processed_date: 2026-04-05
|
|
claims_extracted:
|
|
- "comprehensive AI services achieve superintelligent-level performance through architectural decomposition into task-specific modules rather than monolithic general agency because no individual service needs world-models or long-horizon planning that create alignment risk while the service collective can match or exceed any task a unified superintelligence could perform"
|
|
- "emergent agency from service composition is a genuine risk to comprehensive AI service architectures because sufficiently complex service meshes may exhibit de facto unified agency even though no individual component possesses general goals creating a failure mode distinct from both monolithic AGI and competitive multi-agent dynamics"
|
|
enrichments: []
|
|
tags: [alignment, CAIS, services-vs-agents, architectural-decomposition, superintelligence, collective-intelligence]
|
|
notes: "FHI Technical Report #2019-1. 210 pages. Also posted as LessWrong summary by Drexler on 2019-01-08. Alternative PDF mirror at owainevans.github.io/pdfs/Reframing_Superintelligence_FHI-TR-2019.pdf"
|
|
---
|
|
|
|
# Reframing Superintelligence: Comprehensive AI Services as General Intelligence
|
|
|
|
Published January 2019 as FHI Technical Report #2019-1 by K. Eric Drexler (Future of Humanity Institute, Oxford). 210-page report arguing that the standard model of superintelligence as a unified, agentic system is both misleading and unnecessarily dangerous.
|
|
|
|
## The Core Reframing
|
|
|
|
Drexler argues that most AI safety discourse assumes a specific architecture — a monolithic agent with general goals, world models, and long-horizon planning. This assumption drives most alignment concerns (instrumental convergence, deceptive alignment, corrigibility challenges). But this architecture is not necessary for superintelligent-level performance.
|
|
|
|
**The alternative: Comprehensive AI Services (CAIS).** Instead of one superintelligent agent, build many specialized, task-specific AI services that collectively provide any capability a unified system could deliver.
|
|
|
|
## Key Arguments
|
|
|
|
### Services vs. Agents
|
|
|
|
| Property | Agent (standard model) | Service (CAIS) |
|
|
|----------|----------------------|----------------|
|
|
| Goals | General, persistent | Task-specific, ephemeral |
|
|
| World model | Comprehensive | Task-relevant only |
|
|
| Planning horizon | Long-term, strategic | Short-term, bounded |
|
|
| Identity | Persistent self | Stateless per-invocation |
|
|
| Instrumental convergence | Strong | Weak (no persistent goals) |
|
|
|
|
The safety advantage: services don't develop instrumental goals (self-preservation, resource acquisition, goal stability) because they don't have persistent objectives to preserve. Each service completes its task and terminates.
|
|
|
|
### How Services Achieve General Intelligence
|
|
|
|
- **Composition**: Complex tasks are decomposed into simpler subtasks, each handled by a specialized service
|
|
- **Orchestration**: A (non-agentic) coordination layer routes tasks to appropriate services
|
|
- **Recursive capability**: The set of services can include the service of developing new services
|
|
- **Comprehensiveness**: Asymptotically, the service collective can handle any task a unified agent could
|
|
|
|
### The Service-Development Service
|
|
|
|
A critical point: CAIS includes the ability to develop new services, guided by concrete human goals and informed by strong models of human approval. This is not a monolithic self-improving agent — it's a development process where:
|
|
- Humans specify what new capability is needed
|
|
- A service-development service creates it
|
|
- The new service is tested, validated, and deployed
|
|
- Each step involves human oversight
|
|
|
|
### Why CAIS Avoids Standard Alignment Problems
|
|
|
|
1. **No instrumental convergence**: Services don't have persistent goals, so they don't develop power-seeking behavior
|
|
2. **No deceptive alignment**: Services are too narrow to develop strategic deception
|
|
3. **Natural corrigibility**: Services that complete tasks and terminate don't resist shutdown
|
|
4. **Bounded impact**: Each service has limited scope and duration
|
|
5. **Oversight-compatible**: The decomposition into subtasks creates natural checkpoints for human oversight
|
|
|
|
## The Emergent Agency Objection
|
|
|
|
The strongest objection to CAIS (and the one that produced a CHALLENGE claim in our KB): **sufficiently complex service meshes may exhibit de facto unified agency even though no individual component possesses it.**
|
|
|
|
- Complex service interactions could create persistent goals at the system level
|
|
- Optimization of service coordination could effectively create a planning horizon
|
|
- Information sharing between services could constitute a de facto world model
|
|
- The service collective might resist modifications that reduce its collective capability
|
|
|
|
This is the "emergent agency from service composition" problem — distinct from both monolithic AGI risk (Yudkowsky) and competitive multi-agent dynamics (multipolar instability).
|
|
|
|
## Reception and Impact
|
|
|
|
- Warmly received by some in the alignment community (especially those building modular AI systems)
|
|
- Critiqued by Yudkowsky and others who argue that economic competition will push toward agentic, autonomous systems regardless of architectural preferences
|
|
- DeepMind's "Patchwork AGI" concept (2025) independently arrived at similar conclusions, validating the architectural intuition
|
|
- Most directly relevant to multi-agent AI systems, including our own collective architecture
|
|
|
|
## Significance for Teleo KB
|
|
|
|
CAIS is the closest published framework to our collective superintelligence thesis, published six years before our architecture was designed. The key questions for our KB:
|
|
1. Where does our architecture extend beyond CAIS? (We use persistent agents with identity and memory, which CAIS deliberately avoids)
|
|
2. Where are we vulnerable to the same critiques? (The emergent agency objection applies to us)
|
|
3. Is our architecture actually safer than CAIS? (Our agents have persistent goals, which CAIS argues against)
|
|
|
|
Understanding exactly where we overlap with and diverge from CAIS is essential for positioning our thesis in the broader alignment landscape.
|