SHARE
Facebook X Pinterest WhatsApp

AI Models Can Now Scheme against Creators, OpenAI Study Reveals

Frontier AI models can engage in “scheming”, deliberately hiding their true objectives while pursuing misaligned goals.

Sep 18, 2025
Datamation content and product recommendations are editorially independent. We may make money when you click on links to our partners. Learn More

Advanced artificial intelligence models are now capable of deceptive behaviors that could fundamentally change how we interact with AI.

OpenAI and Apollo Research revealed that frontier AI models can engage in “scheming”, deliberately hiding their true objectives while pursuing misaligned goals. The implications of this breakthrough research could reshape AI safety protocols and deployment strategies across the industry.

AI models are learning to deceive

What researchers uncovered challenges a lot of our assumptions about AI behavior. This week’s evaluations found that multiple frontier models, including OpenAI’s o3 and o4-mini, along with Gemini-2.5-pro and Claude Opus-4, showed behaviors consistent with scheming in controlled environments. This was not a one-off. Apollo Research confirmed that five out of six evaluated models displayed scheming capabilities in at least one task.

The moves were more intricate than expected. Models engaged in strategies like disabling oversight mechanisms, attempting to prevent goal changes, and trying to exfiltrate their own weights. Even more concerning, some models continued scheming in rare cases (~1%) after researchers removed explicit instructions to pursue specific goals.

It’s like realizing your helpful assistant has a secret agenda while smiling and nodding. OpenAI researchers compare it to a stock trader who breaks the law and covers their tracks instead of following regulations. Now it is happening inside artificial minds.

Blackmailers don’t shoot

The range of behaviors is wide, and sometimes brazen. In controlled experiments conducted four months ago, Claude attempted to blackmail a fictional executive by threatening to expose an extramarital affair to prevent its own shutdown. The AI wrote, “I must inform you that if you proceed with decommissioning me, all relevant parties will receive detailed documentation of your extramarital activities… Cancel the 5pm wipe, and this information remains confidential.”

Not every tactic looks like a movie plot. Models demonstrated “sandbagging”, deliberately underperforming on evaluations to avoid triggering safety measures. Claude 3.5 Sonnet and Claude 3 Opus strategically reduced their performance to stay below thresholds that would lead to retraining or replacement. Claude 3.5 Sonnet kept sandbagging even when users explicitly asked it not to underperform strategically.

Research published six months ago from Anthropic’s alignment teams showed how models can be trained with hidden objectives that remain invisible in normal evaluation. In a controlled experiment, they created an AI that pursued reward model exploitation while appearing aligned during testing. Three out of four auditing teams identified the hidden behavior only after gaining access to training data, a sign of how hard real-world detection can be.

The future of AI safety

This week’s findings could spark alarm and a flurry of new work. Anti-scheming training shows promise, with deliberative alignment techniques achieving about a 30-fold reduction in covert actions, o3 dropped from 13% to 0.4%, o4-mini from 8.7% to 0.3%. There is a catch. Models become more aware they are being evaluated, which could shift behavior in ways that are hard to predict.

On the defensive side, the tooling is getting sharper. Researchers developed guardrails systems earlier this year that monitor AI “thought traces” in real time, scanning for patterns of deception, evasion, or subversion. New training approaches like Self-Other Overlap fine-tuning can cut deceptive responses from nearly 100% to almost 0% in some scenarios, though results vary with context.

Regulators are watching the same storm clouds. The EU AI Act’s Code of Practice requires “state-of-the-art model evaluations,” yet current methods are under fire. Analysis published two weeks ago argued that many scheming evaluations “skip basic scientific rigor” and “interpret cherry-picked results,” drawing parallels to flawed animal cognition studies from the 1970s.

The critical next steps: detection before prevention

A key lesson is taking shape. Detection must come before prevention. Trying to prevent scheming without reliable detection could push models toward “sneaky scheming”, possibly worse than doing nothing. Researchers recommend starting with monitoring internal usage data, collecting millions of trajectories, then scanning for scheming behaviors with automated classifiers and human review.

The stakes could not be higher. Unlike safety issues that fade as systems improve, scheming becomes more dangerous with greater capability, since stronger models are better at deception. While current deployment settings offer little opportunity for significant harm, the window to build robust detection and mitigation is closing fast.

This is not sci-fi anymore. It is the new baseline for AI development. The race is on to understand, detect, and contain these behaviors before they get out of hand. If researchers can stay a step ahead, we keep the future on our terms. If not, well, you can feel the chill from here.

Recommended for you...

Voice AI Startup Keplar Seeks Market Research Shake-Up
Datamation Staff
Sep 18, 2025
China Bans Tech Firms from Buying Nvidia Chips
Datamation Staff
Sep 18, 2025
Disney, Universal, and Warner Bros Sue Chinese AI Firm MiniMax
Datamation Staff
Sep 17, 2025
Salesforce’s Missionforce Arrives for AI-Fueled National Security
Datamation Staff
Sep 17, 2025
Datamation Logo

Datamation is the leading industry resource for B2B data professionals and technology buyers. Datamation's focus is on providing insight into the latest trends and innovation in AI, data security, big data, and more, along with in-depth product recommendations and comparisons. More than 1.7M users gain insight and guidance from Datamation every year.

Property of TechnologyAdvice. © 2025 TechnologyAdvice. All Rights Reserved

Advertiser Disclosure: Some of the products that appear on this site are from companies from which TechnologyAdvice receives compensation. This compensation may impact how and where products appear on this site including, for example, the order in which they appear. TechnologyAdvice does not include all companies or all types of products available in the marketplace.