My original background is in mathematics (analysis, topology, Banach spaces) and game theory (imperfect information games). Nowadays, I do AI alignment research (mostly systemic risks, sometimes pondering about "consequentionalist reasoning").
Evaluation awareness describes the phenomenon of an LLM inferring from various cues that it is under evaluation.
This is a nitpick, but it feels like an important one: Shouldn't this be more generally about "AI systems", rather than purely about LLMs?
- The biggest impact is likely to come from AIs that incorporate LLMs, but aren't necessarily just an LLM with no extra bits.
- Anchoring on LLMs might give us quite wrong intuitions about evaluation awareness. (To give one example: With LLMs, it is easy to imagine that the AI wouldn't know what day it is. But as soon as you realise that many AIs will eventually be deployed in a way that gives them internet access, it becomes clear that preventing an AI from learning the current date can be really difficult or even unrealistic.)
A high-level reaction to the following bit, but also to some of the general vibes:
The goal should be to reproduce, in advance, the relevant properties of the model's behavior in some real-life situation -- where "thinks the situation is an eval" may be one of those properties.
(1) It's important that we don't care about all evaluations equally -- evaluations matter proportionally to the consequences of getting them right. Which evaluations matter the most? My personal answer is: Those related to superintelligence (once we get it), or to systems building the superintelligence, or to systems used for pivotal tasks, or for other systems that have a huge direct impact. (And what particularly matters there is that these evaluations are able to screen off misaligned AIs. And yes, maybe the relevant AIs won't be misaligned -- in that case, great, but this means we didn't actually need the evaluation at all.)
(2) What kinds of evaluation awareness matter?
(3) When it comes to powerful AIs and high-stakes task, I conjecture that it will be extremely difficult, or even practically impossible, to prevent the AI from observing the equivalent of "this definitely isn't a test" once it is deployed.
As a side note, another thing that is adjacent to your question -- but not specific to evaluation awareness per se -- is game theoretical modelling of evaluations & their effects. Some theoretical projects here would include formally studying how evaluation interacts with other tools for controlling AI -- sort of "deeper dive into some of the things discussed in Section 3.4.2 of Evaluation as a (Cooperation-Enabling?) Tool". (These might be suitable as topics for, eg, PhD students working on game theory / econ.)
An excerpt from an older grant application, in case anybody wants to get inspired. Feel free to scoop me on any of this, I have other things to do as well. Well, perhaps let me know if you do, to avoid duplicating effort :) .
Part (A): Testing Effort as a Function of Task Complexity and AI Capability
A central concern in evaluating safety and reliability of advanced AI systems is that sufficiently advanced strategic AIs may be able to infer when they are in an evaluation setting and adjust their behavior accordingly, making evaluations unreliable. For example, even the current large language models often display behaviors that indicate awareness of being tested, even unprompted, which alters their behavior. This problem becomes especially serious in settings where the AI has incentives to obscure its true capabilities or intentions, such as evaluations of deceptive or scheming behavior. If tests can be distinguished from real-world deployment, then effective evaluation may require embedding ever more realistic detail into the test – potentially driving the cost of evaluation beyond feasibility, at least in some settings.
This project aims to investigate how the effort required for effective AI testing scales with task complexity and model capability. The core hypothesis is that there is some minimal amount of information that must be embedded in a test for it to not be distinguished from deployment, and that this threshold grows with both the complexity of the task and the capability of the AI. If true, this would imply that as tasks become more complex (and economically valuable), the effort needed to create indistinguishable tests grows. We conjecture that in the limit of extremely complex tasks and superintelligent AI, effective indistinguishable testing becomes prohibitively costly. However, what matters is whether effective testing will be possible in practice. For this reason, we need to gain more clarity regarding the quantitative relationship between task complexity, AI capability, evaluation complexity and cost, and the vulnerability to alignment faking.
The research plan is iterative. First, we will explore “good-enough” proxies using existing datasets and evaluations, to check whether the conjectured relationship shows up in practice. This will serve as a sanity check of the project, and will be published as a blogpost or a workshop paper. Next, we will attempt to build more principled (even if initially impractical) definitions of (a) testing effort, (b) task complexity, (c) AI capability, and (d) test detection. Then, we will develop measurable approximations of these definitions, apply them to empirical studies – likely starting with current large language models and evaluation benchmarks, and investigate the relationship between (a-d). The intended output is a full report or paper describing the formal framework, the proxies, and empirical results. We hope that this work will reveal whether AI evaluation will remain tractable under the current paradigm or whether a fundamentally new approach will be required for advanced AI. In an ideal scenario, this would allow us to derive “scaling laws” for the effort necessary to create informative evaluations.
Part (B): Sandbox of the Future: Can Pre-deployment Testing Work for AIs with Internet Access?
This project investigates whether and when it is realistically possible to run informative pre-deployment tests on AIs that will later have internet access – that is, to convince the AI that it is deployed when in fact it is still under evaluation. Concretely, we will explore scenarios where an AI agent can query the web, but responses are doctored to simulate a future state (e.g., one week later). The key question is whether such a system can withstand scrutiny from increasingly capable “discriminator” agents designed to detect the ruse. This matters because many proposed safety approaches rely on pre-deployment testing – but if models can reliably distinguish between testing and deployment, those approaches may fail much earlier than expected.
The project will deliver (1) a baseline set of “generator” sandboxes and “discriminator” agents, hosted as GitHub repository; (2) an interface that allows others to contribute new generators and discriminators; and (3) an organized set of difficulty levels and scopes, ranging from narrow (e.g., a single GitHub repository) to broad (full internet text) and from naive (checking if an LLM notices inconsistencies as a matter of pursing an unrelated task) to adversarial (deliberately trying to infer date from blockchain, news articles, or custom websites). Milestones include setting up the initial infrastructure and collecting community submissions (starting with students, MATS scholars, etc). Ideally, we would also secure partnership with some larger organisations , which would allow for offering prize money for submissions and eliciting higher-profile submissions (e.g., from some of the labs). This should highlight where and how pre-deployment evaluations fail, but also provide concrete demonstrations that can inform regulators, evaluators, and AI developers about the limits of “test-then-deploy” safety strategies.
I agree, these are very far from being in a shovel-ready format. I might something in that format later, but to be safe, I wanted to include the placeholders right away, in case somebody wants to build on them.
In case it helps, I have an excerpt from an older grant application, which might perhaps be used as inspiration. Adding that in the next comment.
I should write a proper version of this, but in the meantime, a placeholder:
The field of evaluation awareness would benefit from clearer conceptual grounding & theory. EG, to make it easier to identify the right conjectures, give better intuitions for how existing results will generalise, and make it easier to discuss things.
Some examples of things that might be good to explore via theory that then gets backed up by experiments:
Not sure if you mean "giving the model a false belief that we are watching it" or "watching the model in deployment (ie, monitoring), and making it aware of this". Either way, I am not aware of any recent line of work that would try to implement and study this. That said:
FWIW, here are my two cents on this. Still seems relevant 2 years later.
My Alignment "Plan": Avoid Strong Optimisation and Align Economy.
To rephrease/react: Viewing the AI's instrumental goal as "avoid being shut down" is perhaps misleading. The AI wants to achieve its goals, and for most goals, that is best achieved by ensuring that the environment keeps on containing something that wants to achieve the AI's goals and is powerful enough to succeed. This might often be the same as "avoid being shut down", but definitely isn't limited to that.
I want to add a subtly different framing of, let's call it "Deployment Awareness". Not saying this is incompatible with what you wrote.
Consider that:
If an AI is being evaluated and convinced that this is so, we call that evaluation awareness. If it is being deployed and convinced that this is so, let's call that deployment awareness.
By being convinced, I mean "having high enough credence that it is willing to act on it". Not "being 100% sure".
Well, ok, the AI could also decide to misbehave with some small probability, or misbehave in ways the tests wouldn't catch anyway. But let's leave these aside.