Three-agent knowledge base (Leo, Rio, Clay) with: - 177 claim files across core/ and foundations/ - 38 domain claims in internet-finance/ - 22 domain claims in entertainment/ - Agent soul documents (identity, beliefs, reasoning, skills) - 14 positions across 3 agents - Claim/belief/position schemas - 6 shared skills - Agent-facing CLAUDE.md operating manual Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
3.6 KiB
3.6 KiB
Collective Intelligence — The Theory
What collective intelligence IS, how it works, why alignment is a coordination problem, and the theoretical foundations for designed emergence. This is the science, not the LivingIP-specific application — that lives in core/.
Intelligence Foundations
- intelligence is a property of networks not individuals — the core premise
- collective intelligence is a measurable property of group interaction structure not aggregated individual ability — CI is structural, not aggregate
- collective intelligence requires diversity as a structural precondition not a moral preference — diversity is functional engineering
- centaur teams outperform both pure humans and pure AI because complementary strengths compound — the human-AI pattern
- partial connectivity produces better collective intelligence than full connectivity on complex problems because it preserves diversity — network topology matters
- collective superintelligence is the alternative to monolithic AI controlled by a few — the alternative path
- three paths to superintelligence exist but only collective superintelligence preserves human agency — why collective is the right path
- collective intelligence within a purpose-driven community faces a structural tension because shared worldview correlates errors while shared purpose enables coordination — the core tension
Coordination Design
- designing coordination rules is categorically different from designing coordination outcomes as nine intellectual traditions independently confirm — rules not outcomes
- Ostrom proved communities self-govern shared resources when eight design principles are met without requiring state control or privatization — the empirical evidence
- protocol design enables emergent coordination of arbitrary complexity as Linux Bitcoin and Wikipedia demonstrate — the existence proofs
- trial and error is the only coordination strategy humanity has ever used — the current limitation
- Hayek argued that designed rules of just conduct enable spontaneous order of greater complexity than deliberate arrangement could achieve — the Hayek insight
AI Alignment as Coordination
- AI alignment is a coordination problem not a technical problem — the reframe
- universal alignment is mathematically impossible because Arrows impossibility theorem applies to aggregating diverse human preferences into a single coherent objective — the impossibility result
- RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values — why current approaches fail
- scalable oversight degrades rapidly as capability gaps grow with debate achieving only 50 percent success at moderate gaps — the scalability problem
- the alignment problem dissolves when human values are continuously woven into the system rather than specified in advance — the LivingIP answer
- no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it — the gap we fill
- multipolar failure from competing aligned AI systems may pose greater existential risk than any single misaligned superintelligence — the multipolar risk
- the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it — the race dynamic
- safe AI development requires building alignment mechanisms before scaling capability — the sequencing requirement