| claim |
ai-alignment |
Major alignment approaches focus on single-model alignment while the hardest problems are inherently collective, creating a massive research gap |
experimental |
Theseus, original analysis |
2026-04-15 |
Collective intelligence architectures are structurally underexplored for alignment despite directly addressing preference diversity value evolution and scalable oversight |
theseus |
structural |
Theseus |
| no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it |
| pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state |
| AI alignment is a coordination problem not a technical problem |
|
| no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it |
| RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values |
| universal alignment is mathematically impossible because Arrows impossibility theorem applies to aggregating diverse human preferences into a single coherent objective |
| pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state |
| no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it |
| democratic alignment assemblies produce constitutions as effective as expert-designed ones while better representing diverse populations |
|