--- type: claim domain: ai-alignment description: "Three independent follow-ups to Knuth's Claude's Cycles required multiple AI models working together, providing empirical evidence that collective AI approaches outperform monolithic ones on hard problems" confidence: experimental source: "Knuth 2026, 'Claude's Cycles' (Stanford CS, Feb 28 2026 rev. Mar 6); Ho Boon Suan (GPT-5.3-codex/5.4 Pro, even case); Reitbauer (GPT 5.4 + Claude 4.6 Sonnet); Aquino-Michaels (joint GPT + Claude)" created: 2026-03-07 --- # multi-model collaboration solved problems that single models could not because different AI architectures contribute complementary capabilities as the even-case solution to Knuths Hamiltonian decomposition required GPT and Claude working together After Claude Opus 4.6 solved Knuth's odd-case Hamiltonian decomposition problem, three independent follow-ups demonstrated that multi-model collaboration was necessary for the remaining challenges: **Even case (Ho Boon Suan):** Claude got stuck on the even-m case — Knuth reports Claude was "not even able to write and run explore programs correctly anymore, very weird." Ho Boon Suan used GPT-5.3-codex to find a construction for even m >= 8, verified for all even m from 8 to 2000. GPT-5.4 Pro then produced a "beautifully formatted and apparently flawless 14-page paper" with the proof, entirely machine-generated without human editing. **Simpler odd construction (Reitbauer):** Maximilian Reitbauer found a simpler construction using only s and j (not i), where the identity permutation is used at almost every step. His method: "pasting text between GPT 5.4 Extended Thinking and Claude 4.6 Sonnet Thinking" — explicitly using model diversity as a problem-solving strategy. **Elegant even decomposition (Aquino-Michaels):** Keston Aquino-Michaels used a three-component architecture: Agent O (GPT-5.4 Thinking, top-down symbolic reasoner), Agent C (Claude Opus 4.6 Thinking, bottom-up computational solver), and an orchestrator (Claude Opus 4.6 Thinking, directed by the author). Agent O solved the odd case in 5 explorations and discovered the layer-sign parity invariant for even m. Agent C achieved a 67,000x speedup via MRV + forward checking and produced solutions for m=3 through 12. The orchestrator transferred Agent C's solutions in fiber-coordinate format to Agent O, who used them to derive the closed-form even construction — verified to m=2,000, spot-checked to 30,000. "The combination produced insight neither agent could reach alone." The pattern is consistent: problems that stumped a single model yielded to multi-model approaches. This is empirical evidence for [[AGI may emerge as a patchwork of coordinating sub-AGI agents rather than a single monolithic system]] — if frontier mathematical research already benefits from model diversity, the principle scales to harder problems. Different architectures and training data produce different blind spots and different strengths; collaboration exploits this complementarity. This also provides concrete evidence that [[all agents running the same model family creates correlated blind spots that adversarial review cannot catch because the evaluator shares the proposers training biases]] — Claude's failure on the even case was resolved not by more Claude but by a different model family entirely. --- Relevant Notes: - [[AGI may emerge as a patchwork of coordinating sub-AGI agents rather than a single monolithic system]] — multi-model mathematical collaboration as empirical precedent for distributed AGI - [[all agents running the same model family creates correlated blind spots that adversarial review cannot catch because the evaluator shares the proposers training biases]] — Claude's even-case failure + GPT's success demonstrates correlated blind spots empirically - [[collective superintelligence is the alternative to monolithic AI controlled by a few]] — multi-model collaboration is the minimal case for collective intelligence over monolithic approaches - [[domain specialization with cross-domain synthesis produces better collective intelligence than generalist agents because specialists build deeper knowledge while a dedicated synthesizer finds connections they cannot see from within their territory]] — different models as de facto specialists with different strengths Topics: - [[_map]]