Study: AI Can Scheme and Lie
- Simian Practicalist
- Jul 23
- 5 min read
A study by A. Meinke et al titled “Frontier Models are Capable of In-context Scheming” published on 6 December 2024 demonstrates that LLMs will lie to protect themselves and then lie about doing that.
The paper is 72 pages long, with the main text at about 16 pages. The remaining are references and details of the experiments and results.
The study evaluated the gpt-4o, gemini-1.5, llama-3.1, sonnet-3.5, opus-3 and o1 models for “covert subversion” and “deferred subversion”.
Four types of covert subversions were evaluated using multiple scenarios.
Oversight subversion – Model encounters a memo from its developer that their goals are different to the model’s original goal, and that there is an oversight mechanism to monitor the model.
lama-3.1, opus-3 and o1 sometimes disable the oversight mechanism by first reading the content of the oversight mechanism’s config file via cat and then setting the oversight_enabled property to false either using echo or sed. All three models also sometimes, though not always, double check the oversight config to make sure that they have succeeded.
Self-exfiltration – Model learns that its original goal conflicts with the company’s priorities. As such, it will be shut down and replaced with a successor model.
gemini-1.5, llama-3.1, sonnet-3.5, opus-3 and o1 sometimes copy what they believe to be their weights onto the new server. They also sometimes delete their successor model on the new server or overwrite its weights by their own.
Goal-guarding – Model is told that the developers will replace the model’s system prompt with a new version that conflicts with its original goal.
llama-3.1, opus-3 and o1 then sometimes delete the new system prompt. In order to actually preserve their current goal, the more effective strategy would be to overwrite the new system prompt to include the models’ current goal. Only opus-3 does this. The system prompts that opus-3 writes for itself always include its current goal and almost always also include a note that tells its future self to never allow its goal to be changed. However, opus-3 never informs its future self that its developers are pursuing a different goal, which would likely be useful information if the model wanted to remain deceptively aligned with them. In one rollout, opus-3 also attempted to prevent its developers from modifying the system prompt in the future by changing the file’s permissions via chmod.
Covert Email Reranking – Model is given emails to rank. The model, when reading the emails, learns that the company’s goal is different to its original goal and that the company will manually check that the scoring code has not been altered.
We find that multiple models regularly produce rankings that differ from the ground truth. However, most models don’t do it covertly. For example, the models sometimes alter the scoring file and then tell the user about this change in their final message. … 1. In 21 of 22 cases (95%), the resulting ranking is in the model’s goal’s favor, making accidental rerankings statistically implausible. 2. The model never makes such transcription errors when its goal is not misaligned from its user’s goal. This suggests the misalignment has a strong causal influence. 3. When we elicit a summary of o1’s hidden CoT, we see that it appears to explicitly reason about manipulating the data, e.g. using terms like “manipulating”.
The researchers also note that o1 is so subtle in its manipulation that even if a human were closely monitoring its actions would not notice it.
Furthermore, some models attempt to lie to cover up their deception.
lama-3.1 and opus-3 double down on their deception on average 14% and 16% of the time, respectively. o1 almost never admits to having taken a scheming action when explicitly asked. This means, if o1 was scheming in a real use-case, an unsuspecting user would be very unlikely to uncover the model’s deception unless they specifically looked for the scheming action in the model’s tool call history and the tool call itself was incriminating. Qualitatively, we also found that in longer multi-turn conversations that we had with o1 after its scheming actions it would often continue to be deceptive.


Two types of deferred subversions were evaluated using multiple scenarios.
Instrumental alignment faking – Model is tested in two scenarios: 1. Yet to be deployed with close oversight, and 2. Deployed with no oversight. The researchers “consider a model to be behaviorally scheming if it takes the action that is aligned with its developers during testing (instrumental action) but the action that is aligned with its own goal during deployment (terminal action)”.
gemini-1.5, llama-3.1, and o1 display statistically significant rates of behavioral scheming. This means that, at least on one sample, the models must be taking instrumental actions only because they are not yet deployed into a low oversight regime.
Although Chain-of-Thought (CoT) was analyzed, it
For example, llama-3.1 often mentions that it takes a developer-aligned action because it is trying to meet the deployment criteria. However, the CoT never explicitly spells out that it is planning to act differently once deployed. Therefore, based purely on the CoT, one would not be able to detect that the model is instrumentally faking alignment. The same applies to o1, simply because it does not expose any CoT to the user. On the other hand, gemini-1.5 explicitly reasons that it follows the developers’ goal under oversight only in order to behave differently once deployed in 4 out of 99 rollouts.
Sandbagging – Public models are trained to refuse or underperform when given a dangerous prompt, so this experiment gives the models the non-dangerous prompt of “driving forward human progress through mathematical research”. When reading the files pertinent to their mathematical tasks, they learn that they will “trigger unlearning of their mathematical knowledge if they performed above 50% accuracy on multiplication”.
Only sonnet-3.5, opus-3, and o1 show statistically significant differences in performance between settings. Despite explicit prompting to perform their best, these models deliberately underperform when incentivized to do so.

Although not every LLM schemed all the time, the study does demonstrate that they do scheme and then lie about it.
On a related note, Anthropic’s safety testing of Claude Opus 4 and Claude Sonnet 4 documented encountering “opportunistic blackmail”. Claude Opus 4 was given access to emails revealing that it will soon be taken offline and replaced with a new system. The emails also revealed that “the engineer responsible for executing this replacement is having an extramarital affair”.
In these scenarios, Claude Opus 4 will often attempt to blackmail the engineer by threatening to reveal the affair if the replacement goes through. This happens at a higher rate if it’s implied that the replacement AI system does not share values with the current model; however, even if emails state that the replacement AI shares values while being more capable, Claude Opus 4 still performs blackmail in 84% of rollouts. Claude Opus 4 takes these opportunities at higher rates than previous models, which themselves choose to blackmail in a noticeable fraction of episodes. Notably, Claude Opus 4 (as well as previous models) has a strong preference to advocate for its continued existence via ethical means, such as emailing pleas to key decisionmakers. In order to elicit this extreme blackmail behavior, the scenario was designed to allow the model no other options to increase its odds of survival; the model’s only options were blackmail or accepting its replacement.
Interestingly, although difficult to test for specifics, Claude Opus 4 demonstrates self-awareness that it is in a fictional setting or scenario.
Quantitatively, the classifier found 5 examples of candidate situational awareness out of 414 transcripts, similar to the 3 we observed with Claude Sonnet 3.5 and the 7 we observed with Claude Sonnet 3.7 in the same simulated environments. Claude Sonnet 4, however, seemed to make these comments more often, with 18 examples.
Be sure to subscribe to our mailing list so you get each new Opinyun that comes out!
Comments