One should assume that AGI, aligned or unaligned, leads to AI takeover. Even if an AI project somehow threaded the needle of creating a superintelligence whose prime directive was obedience to a particular set of human masters, those masters are just a few steps away from becoming posthuman themselves if they wish e.g. for the same level of intelligence as the AI. And if your AI's terminal values include, not just obedience to the wishes of humans (whether that's an autocrat CEO or a world parliament), but rejection of anything that would overthrow human rule, then that's not really an AI-assisted government, it's an AI takeover with a luddite prime directive.
The only kind of "AGI world government" that truly leaves humans in charge, is one in which the AGI deletes itself, after giving the government tools and methods to prevent AGI from appearing ever again.
This note was written as part of a research avenue that I don’t currently plan to pursue further. It’s more like work-in-progress than Forethought’s usual publications, but I’m sharing it as I think some people may find it useful.
Introduction
At some point a company, country, or coalition of countries will successfully build AGI. What happens then?
There are many possibilities, including:
Another possibility, if there’s a large enough intelligence explosion, is that the first project to build AGI organically becomes a de facto world government.
This possibility is worth taking pretty seriously, given the stakes and the fact that an intelligence explosion is fairly likely.
In this note, we’ll briefly outline the argument for expecting the first AGI project to evolve into a world government, and then give some weakly held implications for AGI governance.
We argue that taking this scenario seriously makes it more desirable that:
An important caveat is that we’re just arguing that taking the world government scenario seriously makes these features more desirable than they would otherwise be. We’re not making an argument that they are desirable all things considered (which would require taking many other factors into account).[1]
Why expect the first AGI project to evolve into a world government?
Here’s the basic argument for expecting the first AGI project to become a de facto world government:
In this intelligence explosion scenario, there is a point in time when the first project to build AGI determines what happens next for the world. The project might choose to give power back to other actors (e.g. by open sourcing the models, or giving the model weights to political leadership) — but that would be the project’s choice.
How likely this is to happen depends on the speed, scale and concentration of the intelligence explosion. All other things being equal, the faster the rate of AI capabilities progress, the longer that rapid progress can be sustained (and so the greater the capabilities the resulting superintelligence has), and the greater the extent to which the intelligence explosion can occur without relying on third parties outside of the project, the more powerful the leading AGI project will be compared to the rest of the world. Unfortunately, we don’t currently know how fast, sustained and concentrated any intelligence explosion will be, but given the state of our evidence we cannot rule out that it will be very fast, very sustained, and very concentrated.
It also depends on what type of organisation develops AGI. AGI could be developed by a private company, a single government-led project, or an international consortium of governments. Of these, a private company is least likely to achieve de facto world government status, because their government starts off with far greater hard power than the company, can monitor the activities of the company, and, when it’s clear that the company is becoming extremely powerful, can step in and forcibly take control of the company (or threaten to do so).
The same constraints do not bind government-led AGI projects. However, other countries could potentially maintain the balance of power by making credible threats (of war, or of restricting essential semiconductor manufacturing components) against the leading country and thereby getting access to the model weights. This becomes somewhat less likely to happen if the leading project is a multilateral consortium of governments because such a consortium would have greater hard power, could include the whole of the semiconductor supply chain, and would reduce the number of potentially adversarial countries.
(Weakly held) implications for AGI governance
To the extent that we take the possibility that the first AGI project evolves into world government seriously, we think that the following things become more desirable:
Thanks to many people for comments and discussion.
This article was created by Forethought. See the original on our website.
For example, pushing for AGI development to be government-led might increase the chance that power becomes extremely concentrated (as governments have fewer checks than companies), or that misaligned AI takes over (if you believe that governments would handle this risk less competently than labs).
Bostrom 2014.
How well the project manages to avoid misalignment risk is also an important design feature, but I think AI project designs vary less on this dimension than on how likely they are to become autocracies.
Here’s a very simplified model: at any one time, there’s some chance of the leader of a country having authoritarian impulses, or even being a malevolent actor (like Stalin or Mao). But for democratic countries, at least, this chance is fairly low - let’s say 20%. So if there’s one political leader in charge, we have a 20% chance of that leader trying to make the AGI project autocratic. But if there are political leaders from n countries in charge, where n is the number of countries that would need to coordinate in order to make the coalition autocratic, the chance of autocracy becomes 20%^n. With 4 countries, the chance becomes much less than 1%.
In order to have a flourishing future, we want to have a diversity of moral views, and the ability to make compromises between these different moral views. Having the relevant decision-makers be thoughtful and morally reflective is important, too, but having a diversity of moral views ensures that at least some parties are thoughtful and morally reflective.
Canada, France, Germany, Greece, Italy, Japan, the Netherlands, South Korea, Spain, Sweden, the United Kingdom, Argentina, Brazil, Hungary, India, Indonesia, Israel, Kenya, Mexico, Nigeria, Poland, and South Africa.
The US, Canada, the UK, the Netherlands, Germany, Japan, South Korea, and Australia. These countries are either leading AI developers (US), key security allies of leading AI developers (Canada, UK, Australia) or critical to the semiconductor supply chain (the Netherlands, Germany, Japan, South Korea). The V-DEM Institute categorises all of these countries as liberal or electoral democracies.
In general, the ideal design of a de facto world government is a very hard question, which is another reason to make sure that the initial arrangements are temporary.