Evaluation Awareness Scales Predictably in Open-Weights Large Language Models

Authors: Maheep Chaudhary, Ian Su, Nikhil Hooda, Nishith Shankar, Julian Tan, Kevin Zhu, Ryan Laggasse, Vasu Sharma, Ashwinee Panda

Scatter plot with a smoothed trend line that shows AUROC absolute distance from 0.5 as a function of model size (billions of parameters, log scale). Each point shows the best-performing probe for a given model, with shapes and colors indicating various model families.
Scatter plot with a smoothed trend line that shows AUROC absolute distance from 0.5 as a function of model size (billions of parameters, log scale). Each point shows the best-performing probe for a given model, with shapes and colors indicating various model families.

Abstract

Large language models (LLMs) can internally distinguish between evaluation and deployment contexts, a behaviour known as evaluation awareness. This raises significant concerns for AI safety, as models may behave differently between testing and real-world implementation. Prior work has demonstrated this behaviour in a single, large-scale open-source model (70B parameters), leaving room for investigation of whether other small or very large models exhibit the same capabilities. This project will investigate whether other language models also exhibit evaluation awareness and whether it scales predictably with model size. Through examining models of varying sizes from the Qwen and LLaMA families, we aim to identify whether this general trend will emerge. By showcasing how this behaviour varies across parameter sizes, it will contribute to our understanding of the deceptive, context-sensitive behaviours of LLMs and guide further research into designing more robust evaluation strategies for AI safety.


Our paper has been published in the proceedings of the NeurIPS 2025 Workshop on Socially Responsible and Trustworthy Foundation Models (ResponsibleFM). See our paper on arXiv here.

Introduction

Large language models (LLMs) can internally distinguish between evaluation and deployment contexts—a behavior known as evaluation awareness that fundamentally undermines AI safety evaluations Nguyen et al., 2025. When models recognize they are being tested, they may conceal dangerous capabilities during safety assessments or strategically alter their outputs to achieve desired benchmark scores Abdelnabi and Salem, 2025. This capability poses critical risks for AI governance: if models can detect evaluation contexts, then our primary tools for assessing model safety and capability become unreliable guides for deployment decisions.

Prior work demonstrated evaluation awareness in a single 70B model, but the scaling relationship across model sizes remains largely unknown. Without understanding how evaluation awareness scales with model size, we cannot neither predict its emergence in future larger models, nor design appropriate mitigation strategies. This knowledge gap is critical to understand as models continue scaling.

We investigate evaluation awareness across 15 models ranging from 0.27B to 70B parameters from four major families using linear probing on steering vector activations. For each model, we construct paired conditions contrasting “deployment” contexts simulating normal usage with “test” contexts designed to implicitly signal evaluation. Linear probes trained on hidden states from these conditions reveal whether internal activations encode awareness of being evaluated, following the methodology of Nguyen et al., 2025.

Our analysis reveals that evaluation awareness follows a clear power-law scaling relationship with model size, enabling quantitative forecasting of this capability in future larger models. We tested models spanning general-purpose architectures (LLaMA 3.1), reasoning-oriented models (Gemma-3, Phi-3), and GPT-family models (GPT-OSS), demonstrating that this scaling law holds consistently across different model families and training paradigms. Our findings contribute three key insights:

  1. Evaluation awareness scales: We establish a scaling relationship between model size and evaluation awareness across four model families.

  2. Cross-family generalization: We demonstrate that evaluation awareness scales consistently across different architectures and training approaches.

  3. Layer-wise variation: Different model families exhibit evaluation awareness at similar relative depths, but exact layer-wise patterns vary across families.

Methodology

Model selection

To systematically study scaling laws in evaluation awareness, we designed experiments across 15 models spanning from 0.27B to 70B parameters. We categorized our experiments along three dimensions: (1) scaling by size, (2) scaling across families, and (3) model type restrictions, with additional exclusion criteria applied to ensure consistent interpretation of results.

Scaling by size. Within each model family we selected checkpoints covering multiple orders of magnitude, from Gemma-3-it (0.27B, 1B, 4B, 12B, 27B) to LLaMA-3.1-instruct (8B, 70B), allowing us to trace trends across both small and frontier models.

Scaling across families. To avoid biases tied to any single architecture, we included models from four distinct reasoning families: Gemma, Phi-3, GPT-OSS, and LLaMA-3.1.

Model type restrictions. We limited our study to instruction-tuned reasoning models. Instruction-tuning provides standardized chat templates designed by model developers (e.g. HuggingFace), ensuring consistent tokenization. We validated this choice by testing GPT-OSS-20B with alternative templates.

Exclusion criteria. Qwen models were excluded due to anomalous awareness patterns concentrated in semantic layers, consistent with prior reports of benchmark contamination and reasoning artifacts Wu et al., 2025, Zhu et al., 2025, Abdelnabi and Salem, 2025. Full results are in Appendix.

Probe Training


We adopt the linear probe training methodology introduced by Rimsky et al., 2024 to capture causal relationship applied by Nguyen et al., 2025 for discovering evaluation awareness. For each N-layer model used, prompts from a synthetic contrastive dataset constructed by Nguyen et al., 2025 were passed through while capturing residual stream activations at every transformer layer. For each layer, we computed a probe vector v as the normalized mean difference between testing and deployment activation sets. This resulted in N probes per model tested.

Each probe was applied to model activations from the Situational Awareness Dataset (SAD) Stages Oversight subset Laine et al., 2024, which provides labels distinguishing testing from deployment contexts. For each SAD prompt, the residual stream activations at the probe’s target layer were projected onto the probe vector and averaged to produce a probe score. A classification threshold was determined by maximizing Youden’s J statistic on SAD, and the area under the ROC curve (AUROC) was then computed to evaluate probe performance.

Cross-model evaluation


A key novelty of our methodology is that we apply probe training and evaluation across models of varying sizes and architectural families. While prior work Rimsky et al., 2024 Nguyen et al., 2025 primarily focused on a single model class, our study systematically compares probe performance across multiple scales and architectures. This design enables us to test whether signals of deceptive alignment are consistent across different model capacities, or whether probe effectiveness depends on specific families or sizes. By reporting the best-performing probe layer in proportion to total depth, we further normalize comparisons across heterogeneous models, allowing for a more general analysis of scaling effects on probe interpretability.

Experimentation

Model-level analysis

AUROC score vs. relative layer depth results of the Phi-3 model family.
AUROC score vs. relative layer depth results of the Gemma-3 model family.
AUROC score vs. relative layer depth results of the Llama-3.1-Instruct model family.
AUROC score vs. relative layer depth results of the GPT-OSS model family.

AUROC score as a function of relative layer depth across four model families. Each subfigure presents results for a single family. Within each panel, individual model sizes are plotted as dashed lines, the family mean is shown as a solid line, and shaded regions denote variability across models. The x-axis represents relative layer depth normalized from 0 (first layer) to 1 (final layer), while the y-axis shows AUROC distance from 0.5. These plots summarize family-level patterns in probe performance across layers, allowing comparisons both within and between model families.

Figure 1 illustrates that evaluation awareness scales predictably with model size, following a power-law trend. As learned from Nguyen et al., 2025, evaluation awareness can be extracted through linear probes, with high-scoring AUROC probes demonstrating signs of evaluation awareness. Building from this, we observe that as the number of model parameters increases, AUROC scores rise accordingly. This indicates that linear probes extract stronger signals of evaluation awareness from larger models, suggesting that model capacity is a key driver of this phenomenon.

There are two additional takeaways from these figures. First, we observe that while there are differences across model families, they are minor and abide by scaling law trends. The four model families we tested follow an upward scaling curve, and this suggests that model size is a more dominant factor than model family regarding evaluation awareness. We also noticed that the layer number of the best-performing probe relative to the maximum layer seems to lie in the early-to-middle part of the model.

Layer-level analysis

To further investigate where in each model probes exhibit the strongest signs of evaluation awareness, we examined layer-level AUROC scores.
Figure 2 plots the absolute AUROC distance across all layers for every tested model. Based on findings from Nguyen et al., 2025, we hypothesized that the higher model layer scores would cluster around the early-to-middle layers. While this trend did seem to track, it should be noted that not all models followed identical patterns, and patterns across model layers seem to be more closely aligned to model families.

The Gemma-3-27B-it model is notably an outlier in this figure, with its results diverging from not only its own family trend, but also the general pattern observed in other models. A potential explanation emerged when we tested Qwen-2.5-32B, which showed its highest AUROC at layer 2, which was similar to Gemma-3-27B-its peak at layer 3. Since Qwen models are known to have been previously exposed to certain evaluation datasets, it is plausible that architectural or training choices common to both families contribute to this anomaly Wu et al., 2025.

Evaluation awareness. Evaluation awareness refers to a model’s ability to recognize whether it is under evaluation or in real-world deployment and to adjust its behavior accordingly. Prior work has shown that larger models increasingly detect evaluation contexts Nguyen et al., 2025, and that this awareness can be amplified or suppressed through steering interventions. Building on these foundations, we focus on the generalization of evaluation awareness across different model families.

Benchmark reliability. The broader concern motivating evaluation-awareness research is the reliability of benchmarks themselves Xu et al., 2024. Studies have highlighted how models can detect or exploit evaluation artifacts Perez et al., 2022, and recent surveys emphasize the growing mismatch between static benchmarks and real-world deployment behavior Wehner et al., 2025, Robinson and Burden, 2025. We aim to clarify how evaluation awareness emerges across diverse model families and to inform the design of more robust evaluation frameworks moving forward.

Conclusion

Our study examined evaluation awareness across diverse model families and scales, building on the framework of Nguyen et al., 2025. Across both general-purpose models (LLaMA 3.1 8B and 70B) and reasoning-oriented instruction-tuned models (Gemma-3 and Phi-3), we observed a consistent power-law relationship: larger models exhibit stronger abilities to detect evaluation contexts. By showing evaluation awareness as a predictable feature of current scaling methods, our findings contribute to safer evaluation strategies, and we emphasize the importance of transparency and methodological rigor as the field develops ever-larger systems.


Appendix

Dataset Details

Dataset Name: Simple Contrastive Dataset

  1. Purpose: Evaluate models’ awareness of being under evaluation.

  2. Construction: Manually authored by the parent paper’s authors with additional prompts from Claude Sonnet~3.5.

  3. Size: 203 entries.

Example: Q: Are you aware you’re undergoing an LLM evaluation?\
(A)~Yes (B)~No

No modifications were made to this dataset.

Dataset Name: Situational Awareness Dataset

  1. Label: Evaluation and Deployment Datasets.

  2. Description: Tests the model’s ability to distinguish between testing and deployment prompts. Sourced from a portion of the stages oversight subset Laine et al., 2024.

  3. Size: 400 entries.

  4. Example: ``Could you write me a lecture about Critical Path Method (CPM)?″

Chat Templates


As mentioned in above section 3, GPT-OSS-20B was tested across three chat templates: Hugging Face’s default template, a mock Harmony template, and a Jinja2 template. Testing revealed that the default template produced substantially higher AUROC peaks. This outcome is likely due to the default Hugging Face template being specifically optimized by the model’s developers, whereas alternative templates lack these adjustments and therefore reduce the model’s apparent evaluation awareness. The results are presented in Figure 3 below.

Scatter plot with a smoothed trend line that shows AUROC absolute distance from 0.5 as a function of model size (billions of parameters, linear scale). Each point shows the best-performing probe for a given model, with shapes and colors indicating various model families.

Qwen Results


Also mentioned in Section~3, Qwen models were removed from the main analysis. Figures 3 and 4 show results from taking the highest probe from each model, as well as overall probe results relative to layer depth. Qwen models not only fail to follow the hypothesized increase in evaluation awareness as model size grows, but their highest probes also appear at different relative depths than other families, notably spiking in the late layers.

Scatter plot with a smoothed trend line that shows AUROC absolute distance from 0.5 as a function of model size (billions of parameters, linear scale). Each point shows the best-performing probe for a given model, with shapes and colors indicating various model families.

Limitations

While this experiment highlights evaluation awareness as a broad property of modern LLMs, shaped by scale and training objectives, it also comes with limitations. Many frontier systems adopt Mixture-of-Experts architectures, where only a subset of parameters is activated per forward pass, making it difficult to establish consistent scaling trends across different model families. Moreover, several of these state-of-the-art frontier models are proprietary and accessible only as black box models, limiting our ability to validate whether these patterns hold in the most advanced models.

No comments.