New research reveals concerning patterns in how artificial intelligence responds to international crises, with foundation models showing a troubling bias toward escalation rather than diplomatic solutions. This discovery comes at a critical time when AI systems are increasingly embedded in national security operations, from ChatGPT Gov’s broad governmental deployment to specialized tools like CamoGPT in defense and StateChat in diplomacy.
The big picture: A comprehensive study by the Futures Lab at the Center for Strategic and International Studies (CSIS) and Scale engineers tested AI foundation models against 400 scenarios and over 66,000 question-and-answer pairs to evaluate their crisis management capabilities.
Key findings: The research uncovered significant algorithmic biases in how AI models approach foreign policy decisions and escalation scenarios.
Why this matters: The research suggests AI systems could inadvertently amplify international tensions rather than defuse them, particularly concerning as nations like China develop advanced models like DeepSeek for strategic planning.
Behind the numbers: The study’s extensive testing—covering 400 scenarios and 66,000+ question pairs—provides robust evidence of systematic biases in AI decision-making patterns.
Potential risks: The study identifies several critical concerns about AI’s role in international relations.
The way forward: Researchers recommend specific measures to address these challenges.