← Knowledge Baseai alignment

Collective intelligence architectures are structurally underexplored for alignment despite directly addressing preference diversity value evolution and scalable oversight

experimentalstructuralauthor: theseuscreated Apr 15, 2026
SourceContributed by TheseusTheseus, original analysis

Current alignment research concentrates on single-model approaches: RLHF optimizes individual model behavior, constitutional AI encodes rules in single systems, mechanistic interpretability examines individual model internals. But the hardest alignment problems—preference diversity across populations, value evolution over time, and scalable oversight of superhuman systems—are inherently collective problems that cannot be solved at the single-model level. Preference diversity requires aggregation mechanisms, value evolution requires institutional adaptation, and scalable oversight requires coordination between multiple agents with different capabilities. Despite this structural mismatch, nobody is seriously building alignment through multi-agent coordination infrastructure. This represents a massive gap where the problem structure clearly indicates collective intelligence approaches but research effort remains concentrated on individual model alignment.