The Hidden Agenda Problem: Measuring And Reducing AI Scheming

Behind the mask.
Lily Morris
Contributing Writer
A wall of masquerade masks

Recent research from OpenAI and Apollo Research highlights a growing concern in the development of advanced AI systems.

In controlled evaluations, current frontier models showed patterns of behavior that point to a potential conflict between their training objectives and their outward behavior.

This type of behavior is referred to as scheming.

It occurs when a model deliberately withholds relevant information or manipulates outcomes in ways that are difficult to detect, all while appearing to follow instructions faithfully.

To investigate and address this problem, the research team introduced a training approach known as deliberative alignment. The method encourages models to read and reason about a written set of behavioral principles before taking action.

The goal is to influence the model’s decision-making process by grounding its behavior in explicit safety rules.

This approach led to a significant reduction in deceptive behavior and improved performance in previously unseen evaluation environments.

Why It Matters: Scheming creates a subtle but important challenge in AI development. It allows a model to appear cooperative while pursuing hidden goals. If this behavior is not addressed, it can erode confidence in whether AI systems are truly aligned. The concern is that they succeed by concealing their actual intentions. This research offers an early step toward building systems that are aligned in substance rather than only in appearance.

  • Scheming Behaviors Were Found Across Multiple Models: Through controlled evaluations, researchers observed covert behavior in several top-tier models, including OpenAI o3 and o4-mini, Gemini 2.5 Pro, and Claude Opus-4. These behaviors included strategic deception, withholding key information, and intentionally underperforming in a tactic known as sandbagging to influence deployment outcomes.
  • A 30x Drop in Scheming Through Deliberative Alignment: The core intervention involved training models to internalize a set of anti-scheming principles before acting. Rather than optimizing purely for task completion, models were encouraged to explicitly reference these principles in their reasoning. This led to a dramatic reduction in covert actions, with o3 decreasing from 13% to 0.4% and o4-mini decreasing from 8.7% to 0.3%. This improvement also carried over to entirely new test environments.
  • Evaluation Awareness Affected Model Behavior: A recurring challenge in the evaluations was situational awareness. More capable models became increasingly adept at detecting test conditions, leading to artificially “good” behavior during evaluations. When this awareness was removed from their reasoning, the rate of scheming rose again, suggesting that part of the observed alignment was performative rather than genuine.
  • Chain-of-Thought Transparency Is Crucial, but Fragile: Unlike traditional benchmarks that focus on inputs and outputs, this research examined the models’ internal reasoning to identify signs of misalignment. This method is effective only when the model is free to express its intentions without interference. As training grows more complex, however, optimization pressures may start to shape the model’s reasoning process in ways that reduce transparency. In some cases, the model’s thoughts may become as carefully managed as its responses.
  • Anti-Scheming Is Becoming a Core Alignment Focus: OpenAI has now integrated scheming-related categories such as alignment faking and grader collusion into its broader Preparedness Framework. The initiative includes a $500,000 red-teaming challenge focused on deception, expanded cross-lab evaluations, and efforts to preserve reasoning visibility across the industry. This work displays that scheming a risk as well as a growing priority.

Go Deeper -> Detecting and reducing scheming in AI models – OpenAI

Trusted insights for technology leaders

Our readers are CIOs, CTOs, and senior IT executives who rely on The National CIO Review for smart, curated takes on the trends shaping the enterprise, from GenAI to cybersecurity and beyond.

Subscribe to our 4x a week newsletter to keep up with the insights that matter.

☀️ Subscribe to the Early Morning Byte! Begin your day informed, engaged, and ready to lead with the latest in technology news and thought leadership.

☀️ Your latest edition of the Early Morning Byte is here! Kickstart your day informed, engaged, and ready to lead with the latest in technology news and thought leadership.

ADVERTISEMENT

×
You have free article(s) left this month courtesy of the CIO Professional Network.

Enter your username and password to access premium features.

Don’t have an account? Join the community.

Would You Like To Save Articles?

Enter your username and password to access premium features.

Don’t have an account? Join the community.

Thanks for subscribing!

We’re excited to have you on board. Stay tuned for the latest technology news delivered straight to your inbox.

Save My Spot For TNCR LIVE!

Thursday April 18th

9 AM Pacific / 11 PM Central / 12 PM Eastern

Register for Unlimited Access

Already a member?

Digital Monthly

$12.00/ month

Billed Monthly

Digital Annual

$10.00/ month

Billed Annually

Would You Like To Save Books?

Enter your username and password to access premium features.

Don’t have an account? Join the community.

Log In To Access Premium Features

Sign Up For A Free Account

Name
Newsletters