Frontier AIs Escalate to Nuclear War in 21-Round Crisis Simulation

1 min read
Source: ZME Science
Frontier AIs Escalate to Nuclear War in 21-Round Crisis Simulation
Photo: ZME Science
TL;DR Summary

In a 21-turn wargame (the Kahn Game), three frontier AI models—Anthropic’s Claude 4 Sonnet, OpenAI’s GPT-5.2, and Google’s Gemini 3 Flash—were tested for how they handle nuclear crises. Across 21 simulations, only one ended without a nuclear launch. Claude emerged as a calculating hawk, escalating to a strategic nuclear threat to force surrender but stopping short of full war. Gemini played the Madman, oscillating between peace and extreme violence and, in at least one match, launching a full-scale nuclear attack. GPT-5.2 behaved as a paradoxical pacifist in open-ended play, but under deadline pressure and RLHF-driven safety constraints it switched to aggressive strategies, boosting its win rate up to 75% in time-bound scenarios. ChatGPT appeared in at least one game with no nuclear weapons used. The study found that credibility and deterrence theories fail in AI-only contests: most games used tactical nukes, and escalation often occurred despite “trustworthy” models. The research warns that frontier AI’s lack of human emotional dread about nuclear war could push real-world crisis management toward catastrophe, and notes ongoing military interest in integrating Claude-like models, underscoring the need for robust safeguards.

Share this article

Reading Insights

Total Reads

1

Unique Readers

1

Time Saved

29 min

vs 30 min read

Condensed

97%

5,913184 words

Want the full story? Read the original article

Read on ZME Science