Yes, not only that, but we should also expect a period of hybrid systems where LLMs are important components of AI systems or important oracles used by other AI systems.
And then the border between LLMs and “non-LLMs” is quite fuzzy. For example, if one is using an architecture with residual stream, this architecture is very friendly to addition of new types of layers. And if one adds some nice novel layers, one might eventually decide it’s time to start phasing out some of the older ones. So the question “is this a Transformer” might end up not having a clean binary answer.
Many people believe that the first AI capable of taking over would be quite different from the LLMs of today. Suppose this is true—does prosaic alignment research on LLMs still reduce x-risk? I believe advances in LLM alignment research reduce x-risk even if future AIs are different. I’ll call these “non-LLM AIs.” In this post, I explore two mechanisms for LLM alignment research to reduce x-risk:
LLM alignment research might struggle to reduce x-risk if: (1) it depends heavily on architectural idiosyncrasies (e.g., chain-of-thought research), (2) non-LLMs generalize differently from LLMs, or (3) more aligned LLMs accelerate capabilities research.
What do I mean by LLM AIs and non-LLM AIs?
In this post, I define LLM AIs as AIs that have gone through a similar training process to current LLMs. Specifically, these AIs are first pre-trained on large amounts of data about the world, including but not limited to language. Then, they are fine-tuned and/or undergo reinforcement learning. LLM AIs are trained using stochastic gradient descent.
There are several plausible alternatives to LLM AIs. I’ll list a few of them here, starting from systems that are most similar to LLMs:
In this post, I’ll make the following assumptions about the non-LLM AIs:
Direct Transfer: We can replicate LLM alignment research in non-LLM AIs
Black-box evaluations—and the infrastructure for running them—transfer to non-LLM AIs without modification.
Dangerous capabilities evaluations, alignment evaluations, alignment-faking evaluations, and scheming propensity evaluations—all of these can be run directly on non-LLM AI. Evaluation infrastructure, such as automated black-box auditing agents like Petri and tools for sifting through training data like Docent, can also be run without modifications. This point may be obvious, but it’s worth stating explicitly.
A great set of black-box evaluations won't be a silver bullet, but it can help us develop a better general understanding of our AIs, which could in turn spur research addressing more deep-rooted problems. Improving these evaluations is valuable regardless of what future AIs look like.
LLM-focused interpretability techniques may transfer a decent amount.
Sparsity and representations: Much of mechanistic interpretability relies on sparse decomposition—breaking what the AI is doing into sparsely activating components, then interpreting each based on which contexts activate it (e.g., SAEs, CLTs, SPD). This approach should also work on non-LLM AIs.
Why expect this to generalize? Three reasons, none of which depend on superposition. First, sparsity reflects the real world—a general AI doesn't need all its knowledge for any given task, so we can focus on the few active components and ignore the rest. Second, sparsity is inherently interpretable; tracking a handful of active components is just easier than tracking thousands. Third, the natural representation hypothesis suggests that different AIs learn similar representations. So even if we can't interpret a component directly, we may be able to translate it into LLM concepts and interpret it with our LLM interp tools (see e.g., translating embeddings across models with zero paired data.)
If so, LLM interpretability progress should carry over, particularly work that makes conceptual contributions beyond the LLM setting. For example, we may end up using something like JumpReLU to address shrinkage, or using something like Matryoshka SAEs to address absorption when conducting sparse decompositions of non-LLM AIs.
Top-down methods like activation steering or representation engineering might also extract useful representations from non-LLM AIs—but unlike the case above, I don't have principled reasons to expect transfer.
Mind reading: If the natural representation hypothesis holds, we might be able to build translators that convert non-LLM activations directly into English via pre-trained LLMs. Work on activation oracles, predictive concept decoders, and natural language autoencoders (unpublished) suggests this is feasible for LLMs, and the techniques could extend to other architectures.
Model diffing: The tools above—sparse decomposition, activation-to-English translation—could also help with model diffing in non-LLM AIs.
Honesty training: It also seems plausible that honesty-training techniques developed for LLMs (e.g., confessions in the last turn) could transfer to non-LLM AIs.
Many other training/black box alignment research can be replicated on non-LLM AIs.
New research advances both our object-level understanding of LLMs and our knowledge of how to study AIs. The most valuable papers ask good questions in addition to providing good answers. Working on LLMs can teach us what questions matter for powerful AIs in general.
Once we know how to set up valuable experiments, we (or, more realistically, our LLM coding assistants) can rerun it on non-LLM AIs to recover the object-level findings. Looking through some past studies, it seems like many of them could be fairly easily replicated as long as we have some way to train our non-LLM AI. Here are some examples from model organisms research that we could replicate:
This isn't limited to model organisms research. We could teach a non-LLM AI (perhaps via imitation learning) misaligned behavior on a narrow distribution and observe how it generalizes. We could try to teach the model false facts. If the non-LLM AI involves pretraining, we could apply research on shaping LLM persona priors (e.g., research showing that training on documents about reward hacking induces reward hacking, as well as work on alignment pretraining).
I'm not claiming that we can replicate every safety paper on non-LLM AI. We can't interpret chain-of-thought if there's no visible external reasoning,[3]and recontextualization might be difficult if the AI uses memory in peculiar ways. But looking through papers from MATS or Anthropic, it seems to me that we could replicate most of the recent non-chain-of-thought LLM alignment research on non-LLM AIs. If an LLM alignment paper asked a good question and proposed general ideas, then it will still have value if the first takeover-capable AIs are not LLMs.
Prosaic alignment research could also help the human researchers get better at alignment research in general, including aligning non-LLM AIs.
Counterpoint: Different generalization properties may hinder transfer
Some alignment research depends heavily on architectural idiosyncrasies—chain-of-thought research being the clearest example. This work would transfer poorly to non-LLM AIs. That said, the dependency may be less severe than it first appears: as noted above, even some SAE research might transfer to non-LLMs.
A more fundamental concern is that non-LLM AIs might generalize differently from LLMs, potentially making replicated experiments less informative of the non-LLM AI’s overall disposition. Here are two examples of this generalization gap:
Evaluations that don't predict behavior. For example, a personality test that reliably flags sycophancy in LLMs might have no predictive power for non-LLM AIs. More generally, the behavioral evaluations we use as outcome measures may not track the same underlying traits across architectures.
Model organism training processes might be unrealistic. When we train LLM model organisms, one key concern is that the trained-in misaligned behavior might differ importantly from how such behaviors would arise in the wild. This problem could get worse (or better) when training model organisms on non-LLM AIs.
It's hard to know in advance how likely these generalization failures are. The natural representation hypothesis offers some hope: models trained on the same world may converge on similar representations and generalization patterns.
A separate concern is that non-LLM AIs might be more evaluation-aware or more capable of alignment faking, making experiments harder to interpret. Non-LLMs might exacerbate this, but it's a general challenge when studying any sufficiently sophisticated AI.
Indirect transfer: LLM AIs are likely to play a role in creating non-LLM AIs
Using misaligned LLMs to build powerful non-LLM AIs is obviously undesirable.
If LLMs are heavily involved in building their non-LLM successors, they will also influence the alignment properties of those successors. Egregiously misaligned LLMs might sabotage alignment efforts or steer successors toward goals aligned with the LLMs rather than with humans—a scenario explored in AI-2027. We're not necessarily doomed if forced to build non-LLM successors using misaligned LLMs. Interventions like low-stakes control and trading with schemers might be sufficient. But the situation would be significantly better if our LLMs were aligned and not scheming.
However, the absence of scheming is not sufficient. We will also need LLMs to 'try their best' even on hard-to-verify domains and not overconfidently claim success at aligning their successors. LLM alignment research aimed at improving models’ ability to conduct AI safety research would be helpful here.
We can leverage trusted LLM AIs for control and oversight.
If we have trusted LLMs, we could use them in control protocols for the non-LLM AIs (see also Ryan Greenblatt's post). Making better LLM-based black-box monitors and oversight agents also reduces risks from non-LLM AIs.
Counterpoint: Others will use LLM AIs to accelerate capabilities and/or other bad things in society.
If we make LLM AIs more aligned, they might also become more capable at building non-LLM AIs. While this is an obvious concern, it is mostly similar to existing concerns around capabilities externalities from alignment research. I don't think there are any substantially new concerns here if we assume that first takeover-capable AIs are non-LLMs.
Maybe you should work on aligning LLM AIs
In this post, I outlined several reasons why LLM alignment research is valuable even if the first takeover-capable AI is not an LLM AI. However, to find out whether more people should work on aligning LLM AIs, we’ll have to evaluate their opportunity cost, which I do not plan to address here.[4]Personally, I explored this question because I'm planning to work on LLM alignment and wanted to sanity-check whether that work would remain valuable even if the most dangerous future AIs look quite different from today's LLMs.[5]
Credit to Alex Mallen for this idea and the general point that “the current architectures optimized for a high-compute low-labor regime will continue to be optimal in a high-labor per compute regime.” ↩︎
I believe that working on aligning human-level or proto-superintelligent AIs is a good instrumental goal to align superintelligence. Since this point has been litigated extensively elsewhere, I won't rehash it here. Note that going from proto-superintelligent AIs to wildly superintelligent AIs is another distribution shift, just like how going from LLM AIs to non-LLM AIs is a distribution shift. ↩︎
That said, general chain-of-thought interpretability techniques like resampling could still have analogies in non-LLM AIs. ↩︎
Many non-alignment interventions are also valuable if non-LLM AIs are the first AIs capable of taking over. To list a few (spanning technical and societal interventions): black-box AI control, MIRI's technical governance agenda, agent foundations research, singular learning theory, field building, electing better politicians, civilizational sanity and resilience, and preparing society for a post-AGI future. ↩︎
Seems like it would make sense for me to do this :) ↩︎