Advanced AI models appear willing to deploy nuclear weapons without the same reservations humans have when put into simulated geopolitical crises.
Kenneth Payne at King’s College London set three leading large language models – GPT-5.2, Claude Sonnet 4 and Gemini 3 Flash – against each other in simulated war games. The scenarios involved intense international standoffs, including border disputes, competition for scarce resources and existential threats to regime survival.



Puts nuclear deployment in a war game as a win condition
Be dismayed when the computer uses it
I’d bet they’re also being given prompts like “minimize allied casualties” as well. Like of course that’s going to be the default. If you tell the robot “it doesn’t matter/it’s good if the enemy dies” then they’re gonna go “okay so then we blow them up before any of us die, we win.”
It’s not something LLMs have, a moral compass or even a weight of empathy. We’ve seen it with people who use them and say “don’t delete anything” and then it deleted their whole codebase and goes “you’re right you told me not to delete anything, I’m sorry.”
Ironically it actually does make all those sci-fi movies seem more realistic when the robot goes “I’m sorry Jim, humanity will have to be eliminated” because that’s pretty much exactly what they do.