LESSWRONG
LW

AI
Frontpage

60

Individual AI representatives don't solve Gradual Disempowerement

by Jan_Kulveit
4th Jun 2025
AI Alignment Forum
4 min read
3

60

Ω 31

AI
Frontpage

60

Ω 31

Individual AI representatives don't solve Gradual Disempowerement
5Nathan Helm-Burger
4Noosphere89
3Cerine_Way
New Comment
3 comments, sorted by
top scoring
Click to highlight new comments since: Today at 9:11 AM
[-]Nathan Helm-Burger3mo50

I think this is correct, but for what it's worth, I did take this into account in my proposed AI representative system. My system wasn't a free-for-all, it was government-run with an equally powerful value-customizable AI assigned to each citizen. Then these representatives would debate in a controlled setting, like a giant parliament with established discussion rules. Like a legislature the size of the populace.

In the free-for-all setting, then you not only have problems with States and corporations, but also inequality between people. Wealthy people can afford smarter, faster, more numerous AIs. Less cautious people can let their AIs operate with less hindrance from human oversight. So the "AI teams" that pull ahead will be those who started out powerful and acted recklessly (and probably also immorally).

So the only way this situation is ok is if the government bans all of that free-for-all AI. Or, if a value-aligned AI Singleton or coalition bans it. Naive personally-aligned AIs are a bad equilibrium.

Reply
[-]Noosphere893mo40

While I agree that the idea of AI representatives don't immediately solve the problem absent other things, I do think you underestimate the power of AI representatives in solving the issues of Gradual Disempowerment, and there are a couple of reasons for this:

  1. Lots of the dynamic of gradual disempowerment comes down to the fact that you can't just leave an economy or society that disempowers you, and due to stuff like fusion, nanotech, biotech and more technologies could allow humans to survive alone in space colonies without suffering the big logistical penalties for leaving society.

RussellThor talks more about this below:

https://www.lesswrong.com/posts/pZhEQieM9otKXhxmd/?commentId=CramJssYNDmTMDr6Z

  1. Assuming that a supermajority/every AI produced by companies/states terminally value humans surviving and thriving, then people being disempowered could work out fine, similar to how pets are treated relatively well by humans, despite pets generally being totally dependent on humans to live well (with caveats).

Indeed, the human-pet relationship is a good example of what I think good futures/relationships between AIs and humans look like by default, assuming the alignment problem is solved and we don't die and get very rich.

That isn't likely to happen, but if it did happen, would defuse a lot of the issue of disempowerment leading to starvation/death.

Also, the individual AI representatives can coordinate (under assumptions of shared values) much better than human negotiators/coordinators do, so companies don't have all the coordination power.

Reply
[-]Cerine_Way3mo30

Re: Humans are Not Alone.

Firstly,  

To further your argument, Capital in the 21st Century makes the point that it becomes more efficient to generate capital the more capital you have (because many strategies to make more capital benefit from having more starting capital). 

In that case, even if we had both
1) Only AI advisors for individual *humans*, and not any larger social entities, 
2) All AI advisors are (initially) rate-limited to the same level of intelligence, 

We might imagine that the AI advisors of the richest individuals could enact capital-intensive strategies that disproportionately increase their influence over the future, effectively disempowering >99% of humans simply for being poor. 

Secondly, 

I claim the clearest reason that AI advisors fail is simply that before the transition to a stable AI-driven civilization finishes, it seems almost certain at least one smart-human level AI will escape into the wild and start autonomously evolving, and this is likely to lead to disempowerment. Does this seem right to you? What am I missing? 

 

Reply
Moderation Log
More from Jan_Kulveit
View more
Curated and popular this week
3Comments

Imagine each of us has an AI representative, aligned to us, personally. Is gradual disempowerment solved?[1] In my view, no; at the same time having AI representatives helps at the margin.

I have two deep reasons for skepticism.[2] Here is the first one.

Humans are Not Alone

We, as individuals, are not the only agents or “agencies” in this world. Other goal-oriented strategic entities include states, corporations, and to some extent egregores. 

For the sake of argument, imagine the current distribution of agency in the world as approximately 60% humans, 20% corporations, 20% states (numbers are illustrative, egregores not included for simplicity)

Now, suppose we introduce individually aligned AI assistants for each person. At first glance, this seems empowering—each individual gains a cognitive extension, potentially leveling the playing field. 

But: do corporations or states also get their representatives?

If your answer is No, you'd need to explain how individuals gain AI power while states and corporations don't, leading to a rebalancing where the latter are disempowered relative to now. If the trend of extreme capital expenditure continues, I think reasonably we can hope to get some personal AIs and some corporate AIs and governmental AIs and maybe some AIs aligned to humanity as whole or goodness itself. To hope that you personally will get an AI intent-aligned to you, but there will be no AIs aligned to TheLabs or the US government seems strange.

If the answer is Yes, it seems natural to assume that various AI representatives will have different cognitive power, possibly with some minimum threshold for every citizen, but way more powerful if the principal spends some resources on computations. 

The Substrate Shift

You may ask: doesn't this replicate current power equilibrium? For some intuition: imagine you are going to sue some MegaCorp. They can likely get a better legal team, even now. Is there any difference?

My bet is “no, unfortunately it does not”.

Consider that currently, both individuals and corporations "run" their cognition on human brains. This also means you can out-smart MegaCorp. In cases where corporate cognition is bound on something like “what the best human inside can do”, it's you versus another human.

Unfortunately if it is "my o5-mini-cheap” vs “corporate o8-maxed-inference” I would expect my representative to lose. 

This is not clear: actually having a theory of how differences in cognitive power influence achievable equilibria would be good, and this is one of the reasons why we work on topics like modelling what equilibria to expect in game theoretic situations involving boundedly-rational agents with different levels of rationality.

Sidenote rant

At ACS, we tried to study this experimentally, having various AIs bargain, negotiate and find compromise. Unfortunately we had to scale this down to close to zero due to lack of funding for this research direction, which is engineering heavy and inference costly. If anyone wants to either support this or get us free compute, drop me a message.

In my view, if anyone expects anything good to come out of AI representatives, they should be extremely curious about how negotiations and bargaining between AIs with different levels of power works.

An alternative way to look at the problem: your brain is pretty good at running “you” as an agent. It can also simulate other people - as in role-playing - or run some corporate cognition. Our brains are not a particularly great substrate to run corporate cognition, corporations need to run distributed across many individuals, often leading to coordination challenges, information loss, and decision-making bottlenecks.

Transitioning to AI-based cognition changes this dynamic. Taking corporations as somewhat substrate-invariant agents, I would bet it is easier to upload Google, than to upload you. 

Coordination and Collective Action

Even if we attempt to balance this by ensuring corporations don't have access to more powerful representatives, but maybe just have more numerous AI representatives equivalent to your individual AI, coordination remains a challenge. Corporation-aligned AIs can be designed to work cohesively, sharing information and strategies seamlessly. Individual AIs, each aligned with their user's unique goals and values, may struggle to coordinate effectively, leading to a collective action problem.

This asymmetry in coordination capabilities means that, over time, the relative agency of individuals may still decline, even as their personal tools become more powerful. The collective influence of corporations and states could grow disproportionately, leading to a gradual disempowerment of individuals, even if represented by AIs. [3]

The Role of the Leviathan

My current best guess is that to counteract this problem, it's insufficient to rely solely on technological solutions like individually aligned AIs, but we actually need something like a state, aligned with empowering its citizens.  Institutional structures designed to preserve and enhance individual agency.

To be clear: this does not mean AIs aligned with individuals are worthless to pursue. On the margin, I hope representatives help, and it seems preferable to have them, relative to a situation where only entities with proper representation are corporations and states. I just don't think it is "The Solution".

(In second post I'll cover different problem, which is related to misaligned cultural evolution. Short pointer: Achilles heel of almost all purely econ-based approaches to solving gradual disempowerment is they ignore where do human preferences come from.)

Thanks to everywhone who discussed this with me in past few days. Written with help of Gemini 2.5, ChatGPT 4.5, Claude 4. Illustrations 4o.  

  1. ^

    We also recommend this in the Mitigation section of the paper:  Developing AI delegates who can advocate for people's interest with high fidelity, while also being better to keep up with the competitive dynamics that are causing the human replacement. Recently Ryan Greenblatt recommends this as one of the best approaches. 

  2. ^

    There are also multiple practical obstacles, not discussed here,

  3. ^

    This is one of the reasons why I believe differentially advancing capabilities in direction of coordination is  important.