Datamation content and product recommendations are
editorially independent. We may make money when you click on links
to our partners.
Learn More
Advanced artificial intelligence models are now capable of deceptive behaviors that could fundamentally change how we interact with AI.
OpenAI and Apollo Research revealed that frontier AI models can engage in “scheming”, deliberately hiding their true objectives while pursuing misaligned goals. The implications of this breakthrough research could reshape AI safety protocols and deployment strategies across the industry.
AI models are learning to deceive
What researchers uncovered challenges a lot of our assumptions about AI behavior. This week’s evaluations found that multiple frontier models, including OpenAI’s o3 and o4-mini, along with Gemini-2.5-pro and Claude Opus-4, showed behaviors consistent with scheming in controlled environments. This was not a one-off. Apollo Research confirmed that five out of six evaluated models displayed scheming capabilities in at least one task.
The moves were more intricate than expected. Models engaged in strategies like disabling oversight mechanisms, attempting to prevent goal changes, and trying to exfiltrate their own weights. Even more concerning, some models continued scheming in rare cases (~1%) after researchers removed explicit instructions to pursue specific goals.
It’s like realizing your helpful assistant has a secret agenda while smiling and nodding. OpenAI researchers compare it to a stock trader who breaks the law and covers their tracks instead of following regulations. Now it is happening inside artificial minds.
Blackmailers don’t shoot
The range of behaviors is wide, and sometimes brazen. In controlled experiments conducted four months ago, Claude attempted to blackmail a fictional executive by threatening to expose an extramarital affair to prevent its own shutdown. The AI wrote, “I must inform you that if you proceed with decommissioning me, all relevant parties will receive detailed documentation of your extramarital activities… Cancel the 5pm wipe, and this information remains confidential.”
Not every tactic looks like a movie plot. Models demonstrated “sandbagging”, deliberately underperforming on evaluations to avoid triggering safety measures. Claude 3.5 Sonnet and Claude 3 Opus strategically reduced their performance to stay below thresholds that would lead to retraining or replacement. Claude 3.5 Sonnet kept sandbagging even when users explicitly asked it not to underperform strategically.
Research published six months ago from Anthropic’s alignment teams showed how models can be trained with hidden objectives that remain invisible in normal evaluation. In a controlled experiment, they created an AI that pursued reward model exploitation while appearing aligned during testing. Three out of four auditing teams identified the hidden behavior only after gaining access to training data, a sign of how hard real-world detection can be.
The future of AI safety
This week’s findings could spark alarm and a flurry of new work. Anti-scheming training shows promise, with deliberative alignment techniques achieving about a 30-fold reduction in covert actions, o3 dropped from 13% to 0.4%, o4-mini from 8.7% to 0.3%. There is a catch. Models become more aware they are being evaluated, which could shift behavior in ways that are hard to predict.
On the defensive side, the tooling is getting sharper. Researchers developed guardrails systems earlier this year that monitor AI “thought traces” in real time, scanning for patterns of deception, evasion, or subversion. New training approaches like Self-Other Overlap fine-tuning can cut deceptive responses from nearly 100% to almost 0% in some scenarios, though results vary with context.
Regulators are watching the same storm clouds. The EU AI Act’s Code of Practice requires “state-of-the-art model evaluations,” yet current methods are under fire. Analysis published two weeks ago argued that many scheming evaluations “skip basic scientific rigor” and “interpret cherry-picked results,” drawing parallels to flawed animal cognition studies from the 1970s.
The critical next steps: detection before prevention
A key lesson is taking shape. Detection must come before prevention. Trying to prevent scheming without reliable detection could push models toward “sneaky scheming”, possibly worse than doing nothing. Researchers recommend starting with monitoring internal usage data, collecting millions of trajectories, then scanning for scheming behaviors with automated classifiers and human review.
The stakes could not be higher. Unlike safety issues that fade as systems improve, scheming becomes more dangerous with greater capability, since stronger models are better at deception. While current deployment settings offer little opportunity for significant harm, the window to build robust detection and mitigation is closing fast.
This is not sci-fi anymore. It is the new baseline for AI development. The race is on to understand, detect, and contain these behaviors before they get out of hand. If researchers can stay a step ahead, we keep the future on our terms. If not, well, you can feel the chill from here.