(Thanks to Adam Shimi and Steven Byrnes for feedback.)
(Edit: unsurprisingly, my views have updated some after writing this post. Specifically, I now put much higher credence on solving inner alignment via verification or relaxed adversarial training)
The late 2021 MIRI conversations include discussion about the difficulty of alignment (don't worry, spending hours reading these isn't required for this post). One shared frame which wasn't immediately present in the discussion was a clean delineation of possible approaches to alignment and what they require. I claim that alignment techniques can be usefully understood as deontological, consequentialist, or capability restriction (or a mixture of these) and we'll be going through the challenges associated with constructing sufficiently useful and safe AI using these approaches. I'll also be discussing and motivating a set of issues which I'll refer to as 'the hard problem of AI cognition': we don't have tools for understanding the cognition or intention of AIs produced by current machine learning methods or understanding of how AI decision making depends on training (c.f. inner alignment). I'll define this problem in more detail below, including where this problem does and doesn't show up in alignment. After going through deontological, consequentialist, and capability restriction approaches at a high level and explaining where I think the hard problem of AI cognition is relevant, I'll explain my take on the strategic implications of this analysis and briefly discuss alignment difficulty. We'll be focussing on X-risk, so we won't directly discuss failures which would 'only' result in large losses of life or economic damage. This is primarily a 'first principles' sort of analysis, though I'll be implicitly (and occasionally explicitly) referencing empirical work.
Epistemic status: exploratory. While many of the ideas stated here appear to be widely accepted in the alignment community, I'm a newcomer to the field trying to cover a lot of ground. But everyone keeps telling me to be Very Ambitious and that alignment lacks strategic direction. So, uh, here goes an attempt at that I guess?
The main prerequisite will be the sorts of concepts discussed in AGI safety from first principles.
We'll refer in more detail to:
- The idea of utility maximization and that coherent decisions imply consistent utilities
- Goodhart's law
- The concept of inner alignment
- Power seeking/instrumental convergence
- The current lack of understanding around deep learning generalization and transparency
- The idea of goal-directed behavior and some of the surrounding discussion in the goals vs utility section of the value learning sequence and the deconfusing goal-directedness sequence. I'd tentatively recommend reading this post first before reading those sequences as we'll go through the background and (hopefully) motivate these concepts in a reasonably concise way.
Really, nothing else?
Well, other knowledge or ideas will hopefully be linked as necessary. Also, here are some things which could be helpful to read (though I'd probably recommend reading them after this post):
- A discussion of using an objective framing or a generalization framing of inner alignment
- Model splintering: out-of-distribution behavior
- Reward splintering: model splintering on reward (really utility)
- As mentioned above, the entire value learning sequence (the first two parts seem more important to me) and the deconfusing goal-directedness sequence
- This recent post on consequentialism & corrigibility.
If you'd like to read the late 2021 MIRI conversations (and haven't read them yet), my weakly suggested reading order is:
- This post
- Zvi's gears analysis of AGI intervention
- The conversations themselves
- Interpreting Yudkowsky on Deep vs Shallow Knowledge. Also consider referencing this when the concept of deep knowledge comes up instead of reading it afterwards.
(Of course, this reflects my biases.)
3 Assumptions and focuses
3.1 Sufficient usefulness
I wrote above that we want to make a sufficiently useful and safe AI. What is this sufficiently useful criteria? The criteria is that the alignment tax must be sufficiently small on the capability dimensions we care about. And what is sufficiently small? And which dimensions? Well, I don't think we currently have a good understanding of this (as it requires predicting the future), but 2 typical models are:
- Small enough that alignment can be enforced via governance without too much incentive for defection. This framing is probably more relevant in slow takeoff.
- Small enough that an actor could use a lead in AI capabilities to accomplish a pivotal act safely before unaligned AIs are constructed. Note that under this framing, the 'capability dimensions we care about' are the ones which can be used to cause a pivotal act. If the alignment penalty makes all pivotal acts impossible, then that technique is (approximately) worthless. This framing is more relevant in fast takeoff and the acceptable levels of alignment tax could depend on the capabilities lead.
For the remainder of this post, we'll abstract over these two models as well as different views on takeoff speed, referencing different perspectives as necessary.
But abstracting over everything results in a mess, so we'll make the following assumptions:
Unrestricted, superintelligent, and capable AIs which act like long-term, expected utility maximizers with purely outcome based goals (aka consequentialists) would cause an existential catastrophe if created (mostly by humans) with approaches similar to current ML. When I say 'expected utility maximizers with purely outcome based goals', I mean that the expected utility maximizer cares intrinsically about consequences (consequentialism) instead of about actions (deontology). This is the same (or at least a very similar) idea as goal-directed AI from this post by Rohin (from the value learning sequence discussed above). We'll use the term consequentialism for the remainder of this post. I prefer this term over goal-directed because the word goal isn't quite what we intend, I'm likely to use this concept somewhat differently from how Rohin intended, and the word more nicely sets up the dichotomy with deontology. We'll also go through the distinction between consequentialism and deontology for AI in more detail below. This assumption is due to an inability to construct a human values utility function, an inability to perfectly inner align an agent's utility function, Goodhart's law, and instrumental convergence.
But what about ambitious value learning? Well, my view is that full ambitious value learning is deeply infeasible for reasons discussed in the ambitious value learning part of that same value learning sequence. If partial or slightly less ambitious value learning resulted in purely consequentialist agents, then just this approach won't be sufficient for avoiding existential catastrophe. Of course, it might not result in such agents, for instance the right [meta-preferences][metaprefs] could avoid catastrophe.
Societal and government competence and coordination aren't very high (this informs how hard it is to enforce alignment through governance).
AIs capable of directly causing existential catastrophe (with at least small probability) have to be quite intelligent and capable on at least some dimensions. There are other threat models worth considering, but we won't talk about them much here.
I won't make a case for why these are good assumptions here beyond what I've already said, but would be happy to chat in the comments.
We'll also mostly pretend AIs will be deep neural networks trained with SGD, but I wouldn't be surprised if this post generalizes.
3.3 Capabilities for catastrophe
Note that we'll be focusing on techniques for aligning AIs in the regime where capabilities are sufficient for unaligned AIs to directly cause existential catastrophe (of course, techniques for less capable AIs could generalize). One potential objection to this approach is that in slower takeoff scenarios, the crux of alignment could come down to determining how to get AIs which aren't existentially dangerous to meaningfully speed up alignment research. This would require 'weakly' aligning these 'less' dangerous AIs. Ideally, this would be done without this work on weak alignment contributing to the very likely ongoing research on how to get AIs to do capabilities research. Perhaps this asymmetry could be enacted via the use of secrecy or asymmetrically useful weak alignment strategies. Regardless, we won't be considering this type of scenario here.
4.1 Deontological approaches
Deontological principles are rules for taking actions which aren't based on the consequences of those actions. In other words, deontological principles 'care' about actions instead of their consequences. Under this definition, deontology is the exact complement of consequentialism. Note that some deontological properties can be encoded or modeled using non-pathological utility functions, but for others this might not be possible due to incoherent decisions (it's not important for this post to have intuition about why incoherent decisions are sometimes required). This is because all behavior can be rationalized as expected utility maximization. Confused or wondering about the implications of AIs having these principles? Hopefully the next sections will clarify this, so read on.
4.1.1 Tool AIs are purely deontological AIs
Long-run consequentialism kills us all, so let's now consider AIs which don't care about optimizing their environments. Specifically we'll first consider AIs which have reduced agency: tool AIs. You may have noticed this appears in the deontological approaches section. That's because I claim that tool AIs (as typically described) are just purely deontological AIs. Agency is mostly just a set of capabilities coupled with (long-term) consequentialism. If we wish to remove agency while keeping capabilities, we must remove consequentialism yielding a deontological AI. It may also be possible to reduce agency by removing some capabilities (such as self-modeling), this will be discussed in the section on restriction based approaches. Tool AIs are an extreme version of a deontological approach as they are purely deontological, but they serve as a good exhibit of the weaknesses and safety advantages of deontological AIs as well as the challenges associated with constructing them.
4.1.2 Purely Deontological AI, what is it good for
Given that different people use the term 'tool AI' in somewhat different ways, I will stick with the verbose purely deontological AI from here on.
Note that purely deontological AIs can be capable of modeling consequences, but they don't care about the consequences of their actions. They don't make explicit plans aimed at causing certain outcomes. The fact that purely deontological AIs can be capable of modeling consequences means these AIs can appear agentic. For instance, chess engines like AlphaZero or StockFish are almost certainly purely deontological, but they do appear somewhat agentic. For a more absurd example, consider an AI which only cares about imitating what its actions would be if it were a consequentialist (e.g. a human or some consequentialist AI). For a competent imitator, this is the same as being a consequentialist. So wait! Why have we bothered with defining this class of AIs if it practically includes consequentialists anyway!? Well, this comes down to why the intentions of AIs matter at all.
Intentions determine behavior when out-of-distribution for intelligent and robustly capable agents. By definition, robustly capable agents robustly pursue their goals, so we'll consider the case where an agent will still pursue its goals effectively on an out-of-distribution input. Nearly identical behavior on a specific distribution of tasks from the training environment could be the result of many possible goals (intentions). However, these different goals can be distinguished by different behavior on out-of-distribution inputs. For example, consider some empirical observations of objective robustness failures in which agents 'care' about a feature correlated with the reward and then pursue that feature when out-of-distribution instead of the reward from the original environment. The behavior out-of-distribution reveals more information about the true goals of the agent (in so far as these weak agents can be seen as having goals). There are also more arcane considerations like deceptive alignment. A deceptively aligned agent would have the instrumental goal of doing well at training and not being detected. Its terminal goals could be anything. So clearly this agent (if capable) would perform well in training, but its true goals would result in different behaviour once it determined that it's no longer in training and there isn't need for deception. How would the agent be able to determine this? By detecting the out-of-distribution property that it's free to defect and seize power. In practice this could work via mechanisms such as detecting out of distribution inputs, modeling the world in general, checking if time has passed relative to when it was training, or simply defecting with small probability at any point.
So there can be purely deontological AIs which act like consequentialist agents in their training environments, but we generally expect them to act less like consequentialist agents on out-of-distribution inputs. In general, I would be Very Surprised if a purely deontological AI caused existential catastrophe in a bid for power without having been trained explicitly to do so. It isn't agency which scares us: it's generalizing agency.
In summary, purely deontological AIs are not existentially dangerous by default. They're in the same danger category as a nuclear bomb: they might kill a bunch of people, but you'll have to launch them first.
4.1.3 Constructing sufficiently useful purely deontological AIs
The next question we need to ask is how a sufficiently useful purely deontological AI can be constructed.
How could we train a purely deontological AI to do useful things? We could train to imitate or predict instead of optimizing for outcomes. Perhaps apply some iterative amplification or similar and boom, you've got an tool AI which can do useful things.
Did you catch it?
The error in the above reasoning? Take a second and think through what's wrong before peeking.
Just because an AI is trained to imitate or predict doesn't mean it's guaranteed to be a purely deontological AI!
For instance, consider an AI trained to imitate another AI which is a competent consequentialist. It seems quite plausible that this imitator would itself just become a consequentialist!
More generally, inner alignment is not guaranteed by all training procedures.
To be clear, I don't think this is a common misconception among people working on or thinking about alignment. This was also a very weak trap example. Regardless, it does seem like a potential pitfall, so I thought I would try to strongly push readers away from this mistake.
So, there isn't an obvious way to train a purely deontological AI. In fact, we don't even know how to check if an AI cares about consequences or deontological rules. Inner alignment with current machine learning approaches is hard. We have no physics or even biological style models for understanding the eventual intentions of superintelligent AI produced via such a process. We don't have solid approaches for inspecting the decision making of deep agents. Or a decent understanding of what decision making will result from a specific training process. We don't know why or how deep learning generalizes. And it's unclear if techniques will generalize to higher intelligence and capability regimes. This is the 'the hard problem of AI cognition' which we'll be referencing throughout the post.
Is this just the entire alignment problem? Well no, it doesn't include outer alignment and it's possible that we could solve alignment without solving this issue either via the rather dangerous approach discussed in the upcoming section on restrained AI or if it simply happens to be easy to get certain cognitive properties regardless of a lack of understanding. Things could happen to work without us understanding why they work. I claim that relying on this is a very dangerous approach due to difficulties evaluating alignment, for instance, consider deceptiveness. So, my view is that the ways around the hard problem of AI cognition are dangerous (though perhaps I am missing some approaches) and that it is a difficult crux of alignment. I also think that a decent amount of alignment research isn't sufficiently focused on this problem and that we should be more actively working on it. I'll explain my reasoning and what this could look like in my strategic recommendations below.
Now let's transition back to the subject of purely deontological AI. Despite these issues, there are obvious ways to train deep neural networks which ensure that they will be purely deontological. For instance, consider training a (randomly initialized) model to output the value 1. Clearly such a model isn't going to be a consequentialist or even intelligent (unless you think the inductive biases of SGD are actually Magic). But, if the task in question might involve modeling consequences, the question of how to use current machine learning approaches to produce intelligent, non-consequentialist agents is considerably tricker.
In the superintelligent, highly capable regime, what sorts of training and objectives might produce purely deontological agents (as opposed to agents which are at least partially consequentialists)? Well, we're clearly deep into speculation land, because there isn't even currently public knowledge of how to produce a superintelligent, highly capable AI (and I wouldn't publish it if I knew). However, I would be Very Surprised if training agents based on the consequences of their actions (outcomes) in even modestly complex environments with something resembling modern machine learning (e.g. reinforcement learning) could produce capable, superintelligent, and purely deontological AIs. This is putting aside edge cases or the application of some not currently known technique. I'd also make a similar claim about AIs trained to imitate another consequentialist AI. Note that constructing plans also falls into the category of outcome based training (assuming you care about whether or not those plans work!). Also be careful not to over-generalize my statement: I'm just saying that you wouldn't get purely deontological agents, not that you couldn't get partially deontological agents which we will discuss later. So, this leaves the tasks which are classically associated with tool AIs such as prediction (for instance, language models). We'll refer to these tasks as process based as opposed to outcome based. So which process based tasks actually result in purely deontological AIs? I will hold off on speculating here, though I think the answer to this question would be useful. My understanding is that in this conversation Eliezer Yudkowsky says that he thinks that current machine learning techniques couldn't even produce an intelligent and purely deontological model. There's also some speculation in this post on safety in predictive learning.
See this comment for related Elicit predictions.
Now let's suppose that all process based tasks do in fact result in purely deontological agents and consider if such agents can be sufficiently useful.
I'm not currently aware of any pivotal act which can be achieved using a process based task AI. Pivotal acts likely require careful and superintelligent planning which requires reasoning about consequences. One possible route would be to use process based task AI to radically speed up alignment research. However, research requires a large amount of consequentialist agency which process based task AIs can't do without assistance. So, the use of AI would have to radically speed up alignment research while simultaneously still having humans do the consequentialist component. My best guess is that this bottleneck would result in insufficient research speed improvements particularly given that AI would also likely be used for capability research (depending on the capabilities lead). The analysis of other pivotal acts proceeds similarly.
See this comment for a related Elicit prediction.
If purely deontological AI via process based tasks is the main approach to alignment enforced by governance, the benefits of defection would likely seem large to actors as tools want to be agents.
So overall, my belief is that trying to solve alignment for current ML via using purely deontological AIs is very unlikely to succeed.
4.1.4 Partially Deontological AI, what is it good for
Given the capability weakness of the purely deontological AIs we'd be able to create, perhaps we can tolerate some level of consequentialism, but also instill deontological properties. For instance, perhaps we'd like to instill deontological properties like honesty, obedience, corrigibility, or conservativeness which would override consequentialism in some cases or prohibit certain courses of action.
The next relevant question is how the hell do we instill deontological properties? We're back to the hard problem of AI cognition. For any outcome based environment which rewards deontological properties, there exists an agent which simply models that reward for deontological properties as a desirable consequence in its utility function and achieves full marks. For instance, rewarding honesty could be modeled as 'be honest' or as 'appear honest to the overseer'. Note that in some sense this 'consequence model' is the 'correct' model: by definition, outcome based environments reward consequences. So maybe if you set up your deontological property and train to convergence you get a superintelligent, consequentialist agent which also has the desired deontological property. But maybe not.
A further concern is that it might be easier to instill deontological properties in less intelligent agents. This could result from full blown consequentialism requiring relatively advanced capabilities like self-modeling, predication, and reasoning about counterfactuals. For instance, note that for life on earth intelligence seems to correlate with consequentialism. While even insects can look quite consequentialist from afar, this is really a much less generalizable phenomenon than human consequentialism. This difference in difficulty could also result from the fact that consequentialism is the 'correct' model as discussed earlier. Overall, this issue creates the potential for quite a dangerous situation in which there is a smooth transition between dumb deontologist AIs and more intelligent purely consequentialist AIs which deceptively pretend to have deontological properties. Even if the transition isn't smooth, there is still potential for danger. When dialing up the intelligence knob (params, training time, etc), noticing a transition region between having the deontological properties you want, some alignment failures, and then seemingly getting those properties back again should be cause for alarm.
There's an additional problem with partially deontological AIs which didn't exist with purely deontological AIs. If the deontological principles of an purely deontological AI splinter, the AI remains very unlikely to cause existential catastrophe. It merely will have some other deontological properties potentially making the AI less useful. However, if the deontological properties of a partially deontological AI splintered or were merely somewhat off, but the consequentialist capabilities were retained, then it's possible that consequentialism wouldn't be overridden in important cases and the AI would kill us all. We don't just need to ensure that we get deontological properties: we need to ensure we get the right deontological properties and those properties actually prevent existential catastrophe.
Beyond all of these issues, we also now have to worry about the utility function of the agent with respect to consequences. While sufficient deontological properties could ensure that an AI with the wrong utility function didn't kill us all, it might not be very useful. Assuming the utility function of the AI was 'close enough' to desired, partially deontological AI could certainly be sufficiently useful. They can potentially be just as capable as pure consequentialists. However, there are likely trade-offs between the strength of deontological properties and the capabilities of the agent. Sufficiently strong conservatism results in doing nothing at all.
4.2 Consequentialist approaches
4.2.1 Myopic Agents
Given that we assume that long-term consequentialists would kill us all, what consequentialist approaches are left? Well, consequentialists which don't care about the long run of course! These are typically described as myopic agents. Unfortunately, we currently don't know how to construct myopic agents: simply training agents with myopic reward is insufficient. This is the hard problem of AI cognition once again. However, I'd note that it would likely be easier to create myopic agents than to instill complex deontological values. However, even if we know how to construct myopic agents, they have potential safety issues mostly related to decision theory. Resolving these safety issues also likely requires grappling with the hard problem of AI cognition. Specifically, many issues could be resolved via producing an agent which follows dumb decision theory. My understanding is that getting an AI to use a specific decision theory is likely to be considerably harder than instilling deontological properties. Of course this would depend on the specific deontological property. In summary, just creating myopic agents might be easier than instilling deontological properties, but we might need to control the decision theory of myopic agents which could make the overall difficulty harder.
What about sufficient usefulness? The alignment tax will depend on how limited the time horizon must be to ensure safety. We'll assume that this time horizon maps onto real world time and that myopic agents follow dumb decision theory. Following dumb decision theory implies that the myopic agent will always pick the choice which maximizes utility: superrationality isn't a concern. So, the time horizon sufficient for safety depends on how fast the agent could seize power and the additional utility associated with possessing power. For instance, suppose that the time horizon was several years, the AI could seize power in a few days, and the AI would be able to achieve much more utility over the remaining period than if it didn't seize power. This time horizon clearly results in existential catastrophe. Unfortunately, it's likely to be difficult to predict how much higher utility would be for a given agent if it possessed power (presumably utility increases over time when possessing power due to higher and higher utilisation). So, the time horizon would likely need to be shorter or around the same time as it would take the AI to fully seize power. This is also difficult to predict, but it does seem to imply that enacting a pivotal act with a myopic AI would be difficult. This is because enacting a pivotal act would probably take around the same amount of time as seizing power. However, perhaps the pivotal act could be 'split up' such that a myopic AI achieving high utility on the shorter term consequences of its actions would be sufficient.
Another potential power seeking concern is that the AI might not have time to seek all power but it would have time to seek some power (possibly causing a disaster) and this would result in higher utility. While it might be possible to stop an AI which has only acquired some power, if the AI wasn't stopped, it might keep acquiring some power continuously eventually resulting in existential catastrophe. The trade-off between utilisation and acquisition of power would depend on the discounting of the agent's utility function within the time horizon. This could push the required time horizon down even further.
Given how short the time horizon would have to be, the usefulness analysis of myopic agents proceeds very similarly to that of purely deontological agents executing process based tasks. It would be very difficult to enact a pivotal act and the benefits of defection would likely seem large when trying to enforce alignment via governance. In the context of governance, there is also a slippery slope concern if agents become discontinuously dangerous at some time horizon. Raising the time horizon would initially appear safe. Until it isn't.
In summary, creating safe myopic agents is likely to be harder than instilling deontological properties (due to issues with decision theory) and the alignment tax is also likely to be higher. Overall, I think myopic agency isn't as promising as deontological approaches.
4.2.2 Getting utility closer to alignment
Even though long-term consequentialists would kill us all, there's still value in the ability to produce agents with consequentialist utility functions reasonably close to the desired one, particularly for complex or hard to specify utility functions. This greatly affects the applicability of deontological and restriction based approaches. Additionally, reward models or classifiers for complex or hard to specify behavior could be very useful building blocks for instilling deontological properties. This also implies that knowing how to construct these components (value learning) would be useful for researching how to instill complex deontological properties in general.
I think that (unaligned) corporate and academic research might end up meaningfully contributing to value learning which reduces the marginal benefit of aligned work in this space. This reasoning is less applicable in scenarios where commercial timelines and/or takeoff are much faster than corporate leadership expects.
4.3 Restrained AIs
The approaches to safety we've seen so far encounter the hard problem of AI cognition (or require getting lucky). Is there another way? Well, we could allow for fully blown, somewhat misaligned, long-term consequentialist AI and restrain it in various ways.
For one example of a restrained AI, consider an AI which operates as a long-term goal achieving consequentialist in some sort of constrained environment (e.g. theorem proving) without even knowing about the real world. Or consider AIs (hopefully) made safer by somehow removing their ability to model humans or another part of the world. Both of these AIs have had their capabilities handicapped to (hopefully) reduce danger. Additionally, the classic 'boxed' AI is an example of this approach and the typically proposed issues with this approach generalize to restrained AIs.
A restrained AI could theoretically not be intent aligned while also being a highly capable consequentialist agent, somewhat impact aligned, and safe. However, any situation where the AI is aware of the restrictions placed on it will set up a conflict. The classic 'physically boxed' AI which is aware of the external world will manipulate and fight for freedom. Similarly, a 'knowledge boxed' AI will fight for missing knowledge if it's aware of the absence. This also applies to naively crafting AIs which are restricted by not knowing their utility function.
My current view is that it would be very difficult to prevent an AI which is aware of considerable parts of the real world from becoming aware of restrictions placed on it. Further, situating a self-modeling AI in an environment or task where it could encounter or reinvent the idea of simulation could result in the AI hypothesizing simulation and proceeding accordingly.
One possible route around these issues is making an AI which is incapable of understanding or using certain types of knowledge. Or similarly, preventing an AI from modeling itself. But now we've bumped back into the hard problem of AI cognition.
In summary, we could either accept that the AI will become aware of restrictions and prepare for conflict or we could avoid conflict by tangling with the hard problem of AI cognition or by severely limiting the possible tasks and environments to prevent the idea of simulation from being considered. Either way, it seems like purely restriction based approaches require a very large alignment tax or embracing quite a bit of danger. My current (very uncertain) view is that the 'prepare for conflict' approach is a Very Bad Idea, so we'll rule it out when speculating about alignment tax. We'll reconsider this when contemplating combining approaches to alignment.
From the pivotal act perspective, it seems very difficult to perform such an act with an AI which has sufficiently restricted knowledge or ability to perform cognition involving the real world. However, the space of possible restrictions is large, so perhaps there are achievable ways to limit the cognition of AIs in order to ensure (or at least increase the probability of) safety while still being able to accomplish a pivotal act. I'm not currently aware of any promising directions.
The enforcement side looks similarly poor. Actors would likely perceive very strong incentives from increasing the capabilities of AIs.
5 Combining approaches
First of all, it's worth noting that restrictions could (slightly) reduce the possibility of catastrophe when testing other approaches in high intelligence regimes. For this to be useful at all, it would require that these restrictions are at least slightly robust to maliciousness. Additionally, this approach is easily fooled by deceptiveness. So, approaches which reduce deceptiveness such as myopia or deontological honesty synergize well with using restrictions for testing. Even AIs which weakly value honesty might reveal malicious intentions if the AI estimates a low probability of successful power seeking and decides not to engage in acausal trade.
It would be possible to combine myopic agency with some deontological properties as an approach for safety. I'm not currently aware of any synergies or anti-synergies; it seems like both approaches would impose some alignment tax and potentially offer different safety properties.
At this time, I don't believe I have any other non-trivial insight into combining the 3 approaches discussed in this post.
6 Strategic recommendations
I was broadly in favor of prosaic alignment work before starting to write this post and I remain so. However, I have updated very slightly against prosaic alignment due to a considerable increase in my estimation of alignment difficulty. See the section on alignment difficulty for more details. My recommendations will focus on prosaic alignment, though I won't make the case for prosaic alignment here.
My recommendations here are hopefully reasonably robust to differing views in takeoff speeds and modest changes in timelines. However, conditioning on very short (<5 years) or somewhat long (>40 years) timelines would probably change the analysis considerably.
While I'm in favor of working on aligning prosaic AI, I think we should actively try to discover new facts about intelligence. Specifically, I think that the alignment community is working too little on the hard problem of AI cognition. I'll propose an idea for a project on this topic and describe how I think the research focuses of the alignment community should change including examples of over and under rated topics, but first I think it's illustrative to go through an example of an existing project which I think is worthwhile and specific examples of additional sub-experiments I think are particularly valuable to conduct.
6.1 Redwood Research's Project
An example of work on instilling deontological properties is Redwood Research's project on getting a language model to never describe someone getting injured. It's probably a good idea to be familiar with this project before reading the rest of this section.
(Disclaimer: I'm TAing for the bootcamp Redwood Research is running and also interested in working there. That said, I think that my inside view differs considerably from the consensus/typical view at Redwood Research.)
As of the time when this post was written, this project has just involved training a classifier for the predicate 'the completion describes injury'. But, the described plan is to use this classifier to train a language model which never violates this predicate. If totally successful, this would be a deontological property of the model. More generally, training classifiers to always identify bad behavior is useful preliminary work for instilling deontological properties.
I think a particularly useful place to extend this project is in trying to understand the changes in cognition and resulting generalization properties of various ways of training models to avoid violating the predicate. Here are some example experiments which I think get at this notion:
- Use this classifier to train a language model from scratch which never violates this predicate, but only train on a dataset which isn't very diverse (for instance, only one genre of fanfiction). Does the property generalize to out-of-distribution inputs? What about adversarially crafted inputs? How does this vary as the training set is made more diverse?
- Instead of training from scratch, instead use a pretrained language model (ideally pretrained or finetuned on this fanfiction dataset) and run the same tests described above.
- Try to interpret and understand the difference in model parameters after training the pretrained model never to violate the predicate. It might be useful to penalize the distance in model weights from the original model.
- Repeat the above experiments but for different model sizes or training strategies and analyze which factors are relevant in how the model generalizes or changes its cognition.
Of course, there are likely cheaper and more effective ways to run experiments which get at the same ideas.
6.2 Ideas for some experiments on RL agent decision making
One domain of experimentation I'm particularly interested in is determining how decision making is modeled inside of RL agents. I'll go through an outline of an example project, but it should be straightforward to think of variants or other experiments which get at the same domain.
Construct a complex and randomly generatable environment which requires trading off various different objectives and probabilities of failure. The environment should reward consequentialism and ideally require higher levels of 'intelligence'. Add an obstacle/trap to this environment which happens to never be positive expected utility to pass through due to a large negative reward and never having a sufficiently positive reward accessible only through this obstacle. Train an agent to succeed in this environment and then see if it will pass though this obstacle/trap under any circumstances. Specifically, test if it will pass through if an out-of-distribution environment is constructed for which a positive reward larger than the obstacle penalty is visible behind the obstacle. Additionally, test how much positive reward is sufficient (if any). Hopefully, this would test if the agent is modeling the obstacle deontologically or as a negative reward in its utility function.
Here are some extensions/variants of this project which I think could be worthwhile (depending on results from the base experiment):
- Vary the complexity of the environment, training time, or the model size/architecture and determine if results change.
- Does avoiding sharing parameters between the agent and the value function (for actor-critic methods) change anything?
- Pre-train the agent in the same environment, but with positive reward possible for passing through the obstacle in some cases. Then, train the agent on the actual environment in which it's never a good idea to pass through the obstacle. Does this change results? How do results change as pre-training and final training times are varied?
- Have multiple types of positive reward in the environment. Occasionally, make it appear as though net positive reward can be obtained by passing through the obstacle, but the reward is actually deceptively lower than it appears in only this case and passing through the obstacle is still net negative. Do this using only one of the types of reward. Then, run the out-of-distribution test for passing through the obstacle using the other type of reward.
- Try to understand where and how the utility function is stored in the model as well as build up a human comprehensible understanding of it. Part of this work could be done using techniques like the ones used in the Understanding RL vision paper. Ideally it should be possible to edit the utility function by changing model weights 'by hand'. Can these adjustments be used to change the behavior of the model with respect to the obstacle?
The details of this exact project could probably be considerably refined, but regardless, I think experiments exploring this general idea would be useful.
6.3 How focus should change
It's worth noting that my analysis here is highly speculative. Specifically, imagine everything in this section is prefixed with 'I speculatively think'.
There are a large number of issues or factors which are difficult to model when reasoning strategically about the focuses of the community (at least without spending much more time on analysis). It's somewhat difficult for me to model how aligned individuals working in technical AI safety are distributed. So maybe claiming there should be less focus on a specific topic is basically irrelevant because individuals working on this topic care about something other than existential risk or have very different empirical views. Similarly, I don't have a good model of how much effort is going into various topics or the difficulty of various topics. Perhaps loads of people are spending lots of time working on interpretability work, but there doesn't appear to be much work here merely because the topic is so challenging. I also don't have a good model of the skills of aligned individuals. It's possible that shifting focus in the way I recommend would result in people working in domains for which they have less skill or aptitude which could make shifting net harmful. As such, individuals, teams, and organizations should take their comparative advantage into account: increased comparative advantage in the current topic should push against changing focuses. Despite all of these issues, I still think this analysis has a chance of being worthwhile. If it actually does prove to be decently helpful, I might spend more time doing a more detailed analysis later.
As stated earlier, the community is spending too little time working on the hard problem of AI cognition for deep learning. This includes instilling deontological properties, understanding the decision making likely to result from various training methods/environments, and building techniques for transparency and interpretability, particularly of decision making itself. In practice this understand may look like biological or ecological models: complex and with known exceptions, but (hopefully) still useful enough. Trying to improve some notion of 'worst case performance' could also be important. It's worth noting that developing architectures which make understanding cognition easier could also be very useful (for instance, architectures which use factored cognition). Overall, focus should shift (at the margin) from making models have behavior X to gaining insight into what sorts of changes in cognition occur when making models do X.
In addition to the Redwood Research project mentioned earlier, here are some examples of endorsed research projects/directions which work on the hard problem of AI cognition (perhaps indirectly):
- TruthfulQA and other work on truthfulness (this would depend on the exact work of course, merely improving on benchmarks doesn't imply progress in understanding)
- Visible Thoughts
- Understanding RL Vision
I think effort should be reallocated away from crafting reward models for more complex or harder to write utility functions. For examples of this type of research, consider Learning from human preferences. This includes work on enhancing human feedback, active learning, improving sample efficiency, and other related topics. This is based on the understanding that a large amount of research is being conducted in this area and based on the analysis in this earlier section of the post.
I don't think it's worth spending the time going over a bunch more topics in detail (before I know more about the reaction to this post), so I'll just do a quick and dirty more/less at the margin in the style of how Tyler Cowen does overrated/underrated. This list probably doesn't include many important topics. If you want me to add a topic or discuss a topic in more detail, say so in the comments.
Less at the margin:
- Reward modeling/preference learning (discussed above).
- Multi-agent interaction and agent-human interaction. This seems like a capability which will be developed by default in slow takeoff. In fast takeoff scenarios, alignment could require agent-human interaction, but I think it's more effective to figure out this cooperative alignment strategy and then determine exactly what agent-human interaction is needed. This is as opposed to working on agent-human interaction in general. I would also guess that capable consequentialism generalizes to successful interaction with humans (assuming intentions are aligned).
- Agent foundations. My not very confident nor well supported intuition is that proving useful things or gaining valuable understanding with this approach is sufficiently unlikely that intellectual resources should be directed elsewhere. This depends on the exact work in question, particularly because agent foundations is used to refer to wide diversity of research. For instance, I think that the conceptual work of Stuart Armstrong and John Wentworth is decently likely to be valuable (despite disagreeing with each of their overall strategies). I'll discuss this more below in a list of my thoughts about some research agendas. I'm not really sure how much their work actually counts as agent foundations.
More at the margin:
- Trying to understand why and how deep learning learns and generalizes. I'm sceptical of mathematically rigorous approaches working, but I'd love to be proven wrong. I'm also somewhat concerned about enhancing capabilities, but I (weakly) think that many types of understanding asymmetrically benefit alignment.
- Interpretability, particularly focused on decision making. This is probably only slightly underrated, but I don't see much work in this space.
About the right amount:
- Trying to predict potential issues with various approaches to alignment.
- Forecasting more generally.
- Factored cognition.
In the same spirit, here are some quick summaries of my thoughts on some recently discussed research strategies/agendas.
Technique focused approaches. Endorsed, though it's critical that we build the underlying intuition needed to understand when and where techniques will generalize.
John Wentworth's Research Agenda. Vaguely endorsed. I agree that better frames and abstractions around agency are important and I endorse many of key questions he seems to be interested in addressing. However, I think there are critical steps between deconfusing agency and being able to do tasks like 'Back out the learned objective of a trained neural net, and compare it to the training objective'. I think that messy, deep learning specific models will need to be built for this type of task, and my vague sense is that this would be the bulk of the work. Additionally, I think building the right abstractions around agency for alignment will require applied work with current machine learning aimed at these exact topics (as opposed to purely conceptual work or depending on already conducted empirical work). It's not entirely clear with me if John would disagree about this previous point based on this post.
Of course, I also disagree about aiming for ambitious value learning as the approach for aligning initial superintelligent AIs, but this isn't very important for the agenda.
Stuart Armstrong's Research Agenda. I think the overall strategy is completely intractable (once again, I'd love to be proven wrong), but that work on many components of this strategy is likely to be useful for more practical approaches. For instance, the projects on this list seem quite worthwhile to me.
7 Alignment difficulty
After conditioning on timelines, takeoff, and AI emerging from deep learning, I think that many of the key cruxes of alignment difficulty are related to the hard problem of AI cognition. Specifically, will we actually develop a good understanding of AI cognition? Even if understanding is poor, how difficult is instilling desired deontological properties and inner alignment more generally? How will this difficulty depend on the intelligence of agents?
Over the course of writing this post, I have noticed my views on the difficulty of alignment have shifted to be closer to my model of the views of Eliezer Yudkowsky. Specifically, my views have shifted considerably toward thinking that alignment is more difficult due to high levels of difficulty in instilling deontological properties, particularly in more intelligent agents. Perhaps this is an example of a general phenomenon: first principles contemplation of consequentialism, agency, and intelligence leads to Yudkowskization (similar to carcinization).
This isn't a novel set of issues, but I haven't seen a thorough discussion of how these issues interact with various approaches to alignment and the strategic implications.
Also note that I'm not claiming that it's necessarily difficult to craft an AI with specific intentions or cognition, just that we have no idea how to do so. ↩︎
When I initially wrote this post, I didn't reference these very relevant sequences which is a pretty big literature review fail on my part. I think I might have lightly read the value learning sequence considerably earlier (I had it on my reading list at some point, but I think I might have skipped it instead of reading it due to reading the first post of the sequence, determining that ambitious value learning was obviously a doomed approach and then moving on despite most of the value in the sequence being in the 2nd part).
These sequences go into more detail than I will here on the exact details of goal-directedness and I'm inclined to steer the definitions/abstractions in a slightly different direction.
Strangely, despite these sequences being relevant, I didn't find that reading them resulted in a much rewriting of this post. I believe that my model and understanding of goal-directedness mostly agrees with the outline Rohin proposes and none of the discussion in the deconfusing goal-directedness sequence resulted in changing my broad strokes views about the idea. Perhaps this indicates that I'm quite good at using motivated reasoning to be lazy, that I'm a stubborn fool, or that my reasoning on these issues was able to convergence to a reasonable state after some less direct exposure from other writings and discussion (such as AGI safety from first principles). ↩︎
For instance, the idea of an AI having a 'deontological goal' of predicting something seems well specified to me, but isn't what we intend when discussing goal-directedness. ↩︎
These 'less' dangerous AIs could still plausibly precipitate disasters of various magnitudes, just not existential catastrophes. ↩︎
I haven't seen this actively discussed despite a decent number of people putting quite high probability on slow takeoff. It also doesn't seem like current alignment research is asymmetric in this fashion. Perhaps trying to asymmetrically improve alignment research speed during this period via weak alignment work should receive more attention? ↩︎
There are some edge cases in which my definition of a purely deontological AI doesn't match typical intuition for tool AIs. These cases will come up below. ↩︎
This may differ from how others use the term tool AI. ↩︎
If an agent is unable to pursue goals effectively, it's unlikely to be able to cause an existential catastrophe, so we won't consider the potential safety issue of capability robustness. ↩︎
Defecting probabilistically doesn't depend on out-of-distribution inputs, but it does represent a guess from the agent that it is out-of-distribution in a deeper sense. ↩︎
Note that this depends on how intelligence is defined. ↩︎
There isn't any way to embed prediction widgets in markdown mode and the LessWrong markdown to custom editor converter butchers footnotes. ↩︎
Unless for some reason deontological properties are likely to splinter into consequentialism? ↩︎
I've also seen purely deontological agents referred to as myopic, but here I'll be using myopic just to refer to consequentialists with little to no weight on long-run consequences. ↩︎
For details, please see the previously mentioned post on safety issues. ↩︎
I'm not claiming that theorem proving requires long-term consequentialism, just that a long-term consequentialist could prove theorems. ↩︎
That said, I'm not very hopeful for factored cognition itself; I don't think the approach makes the types of cognition we most care about considerably easier to understand and it likely makes implementing useful things considerably harder. ↩︎
Just to self-promote, the idea that alignment means precisely matching a specific utility function is one of the main things I wrote Reducing Goodhart to address. The point that this is incompatible with humans being physical systems should be reached in the first two posts.
Though I've gotten some good feedback since writing it and am going to do some rewriting (including explaining that "reducing" has multiple meanings), so if you find it hard to muddle through, check back in a month.
Just curious - how much time have you invested in the DL literature vs LW/sequences/safety?
One thing that consistently infuriates me is the extent to which the AI-safety community has invented it's own terminology/onotology that is largely at odds with DL/ML. For example I had to dig deep to discover that 'inner alignment' mostly maps to 'generalization error', and that 'consequentialist agent' mostly maps to model-based RL agent.
It's like the AI safety community is trying to prevent an unaligned Rome from taking over the world, but they are doing all their work in Hebrew instead of Latin.
Powerful AGI will be model-based RL (consequentialist) agents in one form or another - not much uncertainty there. The AI safety community doesn't get to decide the shape of successful agents, so giving up on aligning model-based agents is just . . . giving up.
The utility functions of human children aren't 'perfectly inner aligned' with that of their parents, but human-level alignment would probably be good enough. Don't let perfect be the enemy of the good.
Achieving human-level alignment probably won't be easy, but it's also obviously not impossible. The success of DL in part is success in partial reverse engineering the brain, suggesting future success in reverse engineering human alignment (empathy/altruism).
Nobody likes jargon (well, nobody worth listening to likes jargon) but there's a reason that healthy fields have jargon, and it's because precise communication of ideas within a field is important. "Inner alignment" indeed has some relationship to "generalization error" but they're not exactly the same thing, and we can communicate better by using both terms where appropriate.
If your complaint is lack of good pedagogical materials, fair enough. Good pedagogy often exists, but it's sometimes scattered about. Plus Rob Miles I guess.
"Consequentialist" is a common English word, defined in the dictionary as "choosing actions based on their anticipated consequences" or something. Then the interesting question is "to what extent do different AI algorithms give rise to consequentialist behaviors"? I don't think it's binary, I think it's a continuum. Some algorithms are exceptionally good at estimating the consequences of actions, even OOD, and use those consequences as the exclusive selection criterion; those would be maximally consequentialist. Some algorithms like GPT-3 are not consequentialist at all.
I think I'd disagree with "model-based RL = consequentialist". For example, a model-free RL agent, with a long time horizon, acting in-distribution, does lots of things that look foresighted and strategic, and it does those things because of their likely eventual consequences (as indirectly inferred from past experience). (What is a Q-value if not "anticipated consequences"?) So it seems to me that we should call model-free RL agents "consequentialist" too.
I would say that model-based RL agents do "explicit planning" (whereas model-free ones usually don't). I don't think "agent that does explicit planning" is exactly the same as "consequentialist agent". But they're not totally unrelated either. Explicit planning can make an agent more consequentialist, by helping it estimate consequences better, in a wider variety of circumstances.
(I could be wrong on any of these, this is just my current impression of how people use these terms.)
So I said consequentialist mostly maps to model-based RL because "choosing actions based on their anticipated consequences" is just a literal plain english description of how model-based RL works - with the model-based predictive planning being an implementation of "anticipating consequences".
It's more complicated for model-free RL, in part because with enough diverse training data and regularization various forms of consequentalist/planning systems could potentially develop as viable low complexity solutions.
But effective consequentalist-planning requires significant compute and recursion depth such that it's outside the scope of many simpler model-free systems - and i'm thinking primarily of earlier DM atari agents - so instead they often seem to develop a collection of clever heuristics that work well in most situations, without the ability to explicitly evaluate the long term consequences of specific actions in novel situations - thus more deontological.
Hmm, I would say that DQN "chooses actions based on their anticipated consequences" in that the Q-function incorporates an estimate of anticipated consequences. (Especially with a low discount rate.)
I'm happy to say that model-based RL might be generically better at anticipating consequences (especially in novel circumstances) than model-free RL. Neither is perfect though.
DQN has an implicit plan encoded in the Q-function—i.e., in state S1 action A1 seems good, and that brings us to state S2 where action A2 seems good, etc. … all that stuff together is (IMO) an implicit plan, and such a plan can involve short-term sacrifices for longer-term benefit.
Whereas model-based RL with tree search (for example) has an explicit plan: at timestep T, it has an explicit representation of what it's planning to do at timesteps T+1, T+2, ….
Humans are able to make explicit plans too, although it doesn't look like one-timestep-at-a-time.
Sure you can consider the TD style unrolling in model-free a sort of implicit planning, but it's not really consequentialist in most situations as it can't dynamically explore new relevant expansions of the state tree the way planning can. Or you could consider planning as a dynamic few-shot extension to fast learning/updating the decision function.
Human planning is sometimes explicit timestep by timestep (when playing certain board games for example) when that is what efficient planning demands, but in the more general case human planning uses more complex approximations that more freely jump across spatio-temporal approximation hierarchies.
Prior to several month ago I had mostly read DL/ML literature. But recently I've been reading virtually only alignment literature.
I actually think there are very good reasons the AI-safety community uses different terms (not that we know the right terms/abstractions at the moment). I won't get into a full argument for this, but a few reasons:
Children aren't superintelligent AGIs for which instrumental convergence applies.
For current capability regimes, sure. In the future? Not so clear. Consequentialist is a more general idea.
A genealogical descendancy of agents creating/training agents is isomorphic to a single agent undergoing significant self-modification.
How is 'consequentialist' more general? Do you have a practical example of a consequentalist agent that is different than a general model-based RL agent?
I understand this exchange as Ryan saying "the goals of AGI must be a perfect match to what we want", and Jacob as replying "you can't literally mean perfect, as in not even off by one part per googol, e.g. we bequeath the universe to the next generation despite knowing that they won't share our values", and then Ryan is doubling down "Yes I mean perfect".
If so, I'm with Jacob. For one thing, if we perfectly nail the AGI's motivation in regards to transparency, honesty, corrigibility, helpfulness, keeping humans in the loop, etc., but we mess up other aspects of the AGI's motivation, then the AGI should help us identify and fix the problem. For another thing, we're kinda hazy on what future we want in the first place—I don't think there's an infinitesimal target that we need parts-per-googol accuracy to hit. For yet another thing, I do agree with Jacob that if we consider the fact "We're OK bequeathing the universe to the next generation, even though we don't really know what they'll do with it" (assuming you are in fact on board with that, as I think I am and most people are, although I suppose one could say it's just status quo bias), I think that's a very interesting datapoint worth thinking about, and again hints that there may be approaches that don't require parts-per-googol accuracy.
Normally in this kind of discussion I would be arguing the other side—I do think it will be awfully hard and perhaps impossible to get an AGI to wind up with motivations that are not catastrophically bad for humanity—but "it must be literally perfect" is going too far!
This argument about whether human-level alignment is sufficient is at least a decade old. I suspect one issue is that human inter-human alignment is high variance. The phrase "human-level alignment" could conjure up anything from Ghandi to Hitler, from Bob Ross to Jeffrey Dahmer. If you model that as an adversarial draw, it's pretty bad. As a random draw, it may be better than default unaligned, but still high risk. I tend to view it more optimistically as an optimistic draw, based on reverse engineering human altruism to control/amplify it.
I thought LW/MIRI was generally pessimistic on human-level alignment, but Rob Bensinger said "If we had AGI that were merely as aligned as a human, I think that would immediately eliminate nearly all of the world's existential risk." in this comment which was an update for me.
So as a result I tend to see brain reverse engineering as much higher priority than it otherwise would deserve, for both inspiring artificial empathy/altruism and also shortening the timeframe until uploading.
My take is that the neocortex (and other bits) are running a quasi-general-purpose learning algorithm, and the hypothalamus and brainstem are "steering" that learning algorithm by sending multiple reward signals and other supervisory signals. (The latter are also doing lots of other species-specific instinct stuff that don't interact with the learning algorithms, like regulating heart rate.)
So if we reverse-engineer the neocortex learning algorithm first, before learning anything new about the hypothalamus & brainstem, I think that we'd wind up with a recipe for making an AGI with radically alien motivations, but we still wouldn't know how to make an AGI with human-like empathy / altruism.
I think there's circuitry somewhere in the hypothalamus & brainstem that works in conjunction with the learning algorithms to create social instincts, and I'm strongly in favor of figuring out how those circuits work, and that's one of the things that I'm trying to do myself. :-)
Yes, I concur. The cortex seems to get 90% or more of the attention in neuroscience, but those smaller more ancient central brain structures probably have more of the innate complexity relevant for the learning machinery. That's on my reading list (along with some of your brain articles a friend recommended).
Oh, no, this wasn't what I meant. I just meant that the usage of children as an example was poor because individual children don't have the potential to succesfully seek vast power. There certainly is a level of sufficient alignment of a just consequentialist utility function which looks like 1−ϵ as opposed to 1. I think this ϵ is pretty low, but I reiterate for 'purely long-run consequentialists'. Note that ϵ must exceptionally low for this sort of AI not to seek power (assuming that avoiding power seeking is desired for the utility function, perhaps we are fine with power seeking as we have the desired consequentialist values, whatever those may be, locked in).
Agreed, but these aren't consequentialist properties. At least that isn't how I model them.
I shouldn't have given such a vague response to the child metaphor.