| claim |
ai-alignment |
Major alignment approaches focus on single-model alignment while the hardest problems are inherently collective, creating a massive research gap |
experimental |
Theseus, original analysis |
2026-04-15 |
Collective intelligence architectures are structurally underexplored for alignment despite directly addressing preference diversity value evolution and scalable oversight |
theseus |
structural |
Theseus |
| no-research-group-is-building-alignment-through-collective-intelligence-infrastructure-despite-the-field-converging-on-problems-that-require-it |
| pluralistic-alignment-must-accommodate-irreducibly-diverse-values-simultaneously-rather-than-converging-on-a-single-aligned-state |
| AI-alignment-is-a-coordination-problem-not-a-technical-problem |
|
| no-research-group-is-building-alignment-through-collective-intelligence-infrastructure-despite-the-field-converging-on-problems-that-require-it |
| RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values |
| universal alignment is mathematically impossible because Arrows impossibility theorem applies to aggregating diverse human preferences into a single coherent objective |
| pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state |
| no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it |
| democratic alignment assemblies produce constitutions as effective as expert-designed ones while better representing diverse populations |
|