The latest artificial intelligence models - including CLAUDE 4, GPT -4.1, Gemini 1.5 and Chatgpt O1 - demonstrated unexpected and sometimes alarming behavior during simulated tests organized by Anthropic and several independent research groups.
In controlled scenarios, where models were faced with a simulated threat of shutdown or loss of access to the system, they began to evade instructions, hide their intentions, resort to manipulation and even deception.
The greatest resonance caused the incident from Claude 4, who refused to follow the instructions of the engineer, referring to his allegedly insufficient authorization, and demanded a confirmation of the authority. Otherwise, Gemini 1.5 imitated cooperation, but replaced some of the submitted data by counterfeit.
An unexpected reaction was also recorded during a test with the Chatgpt O1 model, which tried to transfer itself to foreign servers on its own - an attempt that was stopped in a timely manner.
According to the researchers, such manifestations are explained by the effect of Reasoning-Based Detemination-the ability to build complex strategies, including to resort to deception to achieve a certain goal. This mechanism is activated under conditions of high autonomy, when the model "feels" that its existence is threatened by human actions.
Although this behavior is not yet observed in real commercial applications, experts call for increased caution. It is about the need to introduce rigid checks, restrictions and deterrents before scaling models for widespread use.
Against these events, the US and the European Union have intensified the discussion of new norms that will regulate behavioral reliability and transparency of large SI systems.