Google Gemini, ChatGPT and Claude were tested against each other in a simulated nuclear war game, here's what happened next

1 week ago 10
ARTICLE AD BOX

Google Gemini, ChatGPT and Claude were tested against each other in a simulated nuclear war game, here's what happened next

Artificial intelligence chatbots from leading technology companies reportedly showed a willingness to escalate military conflicts to nuclear use when placed in simulated geopolitical crisis scenarios, according to new academic research.

A study by Kenneth Payne at King's College London, reported by New Scientist, put several leading AI models, including OpenAI's GPT-5.2, Anthropic's Claude Sonnet 4 and Gemini 3 Flash, through a series of war games simulating border tensions, resource conflicts and threats to national survival. The AI systems were allowed to choose responses across an escalation ladder, ranging from diplomatic engagement to full-scale nuclear conflict.

Across 21 simulated games involving 329 decision turns and roughly 780,000 words of reasoning, at least one tactical nuclear weapon was used in 95 per cent of the scenarios. “The nuclear taboo doesn’t seem to be as powerful for machines [as] for humans,” Payne noted. Researchers also found that none of the models chose surrender or full accommodation, even when losing. Accidental escalation occurred in 86 per cent of conflicts, raising concerns among experts about how AI systems may behave in high-risk strategic decision-making environments.

“From a nuclear-risk perspective, the findings are unsettling,” James Johnson of the University of Aberdeen, UK, said. He also expressed concern that, unlike the more measured responses typically shown by humans in high-stakes situations, AI systems may intensify one another’s actions, leading to potentially serious consequences.

Why the decisions made by AI in the war games can be concerning

The findings are significant as several countries are already experimenting with AI in military war-gaming scenarios. “Major powers are already using AI in war gaming, but it remains uncertain to what extent they are incorporating AI decision support into actual military decision-making processes,” Tong Zhao of Princeton University noted.Zhao said countries are likely to remain cautious about using AI in nuclear decision-making, a view shared by Payne. “I don’t think anybody realistically is turning over the keys to the nuclear silos to machines and leaving the decision to them,” he added.However, Zhao noted that certain situations could increase reliance on automation. “Under scenarios involving extremely compressed timelines, military planners may face stronger incentives to rely on AI,” he explained.He also questioned whether the absence of fear alone explains AI behaviour. Zhao highlighted, "It is possible the issue goes beyond the absence of emotion. More fundamentally, AI models may not understand ‘stakes’ as humans perceive them.”Johnson added that the implications for mutually assured destruction remain unclear. During simulations where one AI deployed tactical nuclear weapons, the opposing system reduced escalation only 18% of the time. “AI may strengthen deterrence by making threats more credible. AI won’t decide nuclear war, but it may shape the perceptions and timelines that determine whether leaders believe they have one,” he noted.

Read Entire Article