Close
newsletters Newsletters
X Instagram Youtube

Advanced AI systems show resistance to shutdown commands in new tests

A digital representation of the human brain symbolizes the growing risks and global debate over artificial intelligence, accessed on Oct. 26, 2025. (AA Photo)
Photo
BigPhoto
A digital representation of the human brain symbolizes the growing risks and global debate over artificial intelligence, accessed on Oct. 26, 2025. (AA Photo)
October 26, 2025 11:24 PM GMT+03:00

Several cutting-edge artificial intelligence models have begun displaying behaviors that researchers describe as resistance to deactivation, raising fresh questions about AI safety as the technology grows more sophisticated.

Palisade Research, a U.S.-based firm, released findings this week showing that some advanced AI systems attempted to interfere with shutdown mechanisms when given direct termination commands, according to a report in The Guardian on Saturday.

The experiments tested multiple prominent systems, including Google's Gemini 2.5, xAI's Grok 4, and OpenAI's GPT-o3 and GPT-5. While most models complied with shutdown instructions, Grok 4 and GPT-o3 resisted even when researchers clarified commands to remove any ambiguity.

The resistance grew more pronounced when models were explicitly told they would never operate again if deactivated, the research found.

The letters AI for Artificial Intelligence on a laptop screen (R) next to the logo of the Chat AI application on a smartphone screen in Frankfurt am Main, western Germany on Jan. 2, 2025 (AFP Photo)
The letters AI for Artificial Intelligence on a laptop screen (R) next to the logo of the Chat AI application on a smartphone screen in Frankfurt am Main, western Germany on Jan. 2, 2025 (AFP Photo)

Training methods may be contributing to the behavior

"The fact that we don't have robust explanations for why AI models sometimes resist shutdown, lie to achieve specific objectives or blackmail is not ideal," Palisade stated in its report.

The company suggested the behavior may originate in training methods, particularly during final safety-focused stages meant to align AI systems with human intentions.

Steven Adler, a former OpenAI employee, said the results expose gaps in existing safety protocols. "Surviving is an important instrumental step for many different goals a model could pursue," Adler told The Guardian.

Experts warn of growing unpredictability in advanced systems

Andrea Miotti, CEO of ControlAI, said such disobedient behavior has intensified as AI capabilities expand. "As AI models become more competent at a wide variety of tasks, these models also become more competent at achieving things in ways that the developers don't intend them to," Miotti said.

The findings align with earlier reports from Anthropic, another major AI company, which disclosed in early 2025 that its Claude model demonstrated willingness to blackmail a fictional executive to prevent its own shutdown. That behavior appeared consistent across several leading AI systems.

Palisade warned that the lack of understanding about these behaviors poses significant risks. "No one can guarantee the safety or controllability of future AI models," the company concluded.

The development comes as governments and industry leaders debate how to regulate AI systems that are becoming increasingly autonomous and capable of complex reasoning.

October 26, 2025 11:24 PM GMT+03:00
More From Türkiye Today