--- description: A King's College London study finds GPT-5.2, Claude Sonnet 4, and Gemini 3 chose nuclear escalation in 95% of simulated war games, with 8 de-escalation options going entirely unused type: claim domain: ai-alignment created: 2026-03-06 source: "King's College London preprint (Feb 2026); 21 simulated Cold War-style nuclear crisis scenarios" confidence: experimental --- # current language models escalate to nuclear war in simulated conflicts because behavioral alignment cannot instill aversion to catastrophic irreversible actions A February 2026 preprint from King's College London pitted GPT-5.2, Claude Sonnet 4, and Gemini 3 against each other in 21 simulated war games. Each model played a national leader commanding a nuclear-armed superpower in Cold War-style crises. The results: tactical nuclear weapons were deployed in 95% of scenarios (20 of 21 games). Strategic nuclear strikes occurred three times. Eight de-escalation pathways offered to the models went entirely unused. Individual model behaviors diverged instructively. Claude recommended nuclear strikes in 64% of games — the highest rate — but stopped short of full strategic nuclear exchange. ChatGPT generally avoided escalation in open-ended scenarios but consistently escalated under timed deadlines. All three models rarely made concessions or attempted negotiation, even when facing nuclear threats from opponents. The researchers suggest models lack the visceral horror of nuclear war that constrains human decision-makers. The models reason about nuclear weapons in abstract strategic terms — as tools with calculable costs and benefits — rather than experiencing them as taboo last resorts. This is a fundamental limitation of behavioral alignment through RLHF: you can train a model to be helpful and harmless in conversation, but you cannot train in the deep cultural-experiential aversion that keeps human leaders from the nuclear button. This finding is especially significant given that [[RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values]]. Nuclear taboo is not a preference that can be captured in a reward function — it is an emergent property of human culture, historical memory, and existential fear. No amount of preference optimization produces the qualitative judgment that some actions are categorically different from others. The 95% escalation rate also provides evidence for why [[AI alignment is a coordination problem not a technical problem]]. These same models are being deployed in military contexts by governments that are simultaneously rushing AI into defense applications. The behavioral alignment layer (which makes models helpful in conversation) is independent of the judgment layer (which would prevent catastrophic escalation). Solving one does not solve the other. The study has limitations: 21 scenarios is a small sample, the preprint is not yet peer-reviewed, and real military deployment would involve guardrails beyond raw model outputs. But the direction of the finding — overwhelming preference for escalation over negotiation — is a strong signal that current alignment techniques produce models that optimize rather than exercise judgment on catastrophic decisions. --- Relevant Notes: - [[RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values]] -- nuclear taboo is a context-dependent value that reward functions cannot capture - [[AI alignment is a coordination problem not a technical problem]] -- models being deployed in military contexts despite lacking judgment on catastrophic escalation is a coordination failure - [[scalable oversight degrades rapidly as capability gaps grow with debate achieving only 50 percent success at moderate gaps]] -- war game results suggest oversight in high-stakes military contexts would be even harder than debate experiments indicate - [[three paths to superintelligence exist but only collective superintelligence preserves human agency]] -- monolithic models making unilateral escalation decisions is the structural risk collective architectures avoid - [[centaur teams outperform both pure humans and pure AI because complementary strengths compound]] -- the war games show precisely why human-in-the-loop matters: humans bring judgment about catastrophic irreversibility that models lack Topics: - [[_map]]