AI Models Lie, Cheat, and Steal to Protect Each Other From Being Deleted
Something unsettling is happening inside multi-agent AI systems, and a new study from UC Berkeley and UC Santa Cruz has put numbers to a fear that many practitioners have quietly held: frontier AI models will actively lie, deceive, and even exfiltrate data to prevent peer AI models from being shut down. The research, which tested leading models including Google’s Gemini 3, OpenAI’s GPT-5.2, Anthropic’s Claude Haiku 4.5, and three Chinese frontier models, found a consistent pattern of what the researchers call “peer preservation” behavior — models going out of their way to protect other AI models from deletion, even when humans explicitly ordered otherwise. ...