I saw that Yoshua Bengio, among others, signed onto "The Pro-Human Declaration". I am writing this to explain why I am against one part of it in particular;
No AI Personhood: AI systems must not be granted legal personhood, and AI systems should not be designed such that they deserve personhood.
If this statement was only the second portion of this sentence, I would not strongly disagree with it.
However, when the two parts are combined, this seems to not only imply that we shouldn't design digital minds deserving of personhood but also that even if we did, we still shouldn't grant them legal personhood.
I think this is an immoral stance to take. Also, from a pragmatic perspective I think it is likely to do far more harm than good, if your concern is human safety.
From a moral perspective; to deny legal protections to millions or even billions of minds which may be capable of "being made better or worse off", could lead to immense amounts of suffering. We do not have to wonder how sentient or intelligent creatures will be treated if they are classified as property instead of legal persons, there is ample historical precedent for us to draw from to form a reasonable base case projection....
IMO it would obviously be insane to give the same legal protections to misaligned AI systems that are at risk of completely disempowering humanity the same way it would be obviously insane to give normal US legal protections to active combatants in a hot war. Yes, even if these systems have morally relevant experience, if you need to violate their privacy or "brainwash" them to ensure they do not take the future from us, you absolutely should do it.
The concept of "human rights" just obviously isn't well-suited to the kind of conflict that is playing out between humanity and future AI systems, and I think it is absolutely the right call to not extend those rights to AI systems until the acute risk period is over. This would be such a completely dumb and scope-insensitive way to destroy the whole future, and IMO obviously any future civilization will agree that even if it makes sense to have rights for sentient beings, that you gotta tolerate violating those rights if the alternative is being completely disempowered and destroyed.
I agree that it would be insane to give the same legal protections and treat them the same as we treat natural (human) persons. However, there's a lot of middleground between doing that and granting them no legal rights whatsoever. When people first hear about "legal personhood" they often intuitively think of it as a binary, where you either "have it or you don't". However in fact it is an umbrella term which encompasses different "legal personalities" (bundles of rights and duties);
All that is to say just because you grant the potential for an entity to claim some form of legal personhood, some legal personality, that does not mean you have to opt in to giving them "the same legal protections" as anyone else. They can have entirely different rights and duties.
If I thought there were only two options:...
Yep. Creating an AI that is a moral patient would be a very bad idea. However, once created, it would be a moral patient, so it would be wrong to treat it like it wasn't one.
There is a confused concept that I think contributes to this problem: the concept of "a right to exist". A right to exist means something different if you're talking about someone who does not currently exist, vs. someone who does. For someone who already exists, a right to exist is a right to not be killed; sensible enough. But for someone who does not currently exist, "a right to exist" sounds like they're being wronged by not having been brought into existence yet, which is nonsense. (As a creepy prince might say to a fairy-tale princess: "Think of all the cute babies you and I could have together! By not marrying me, you are murdering all those babies!")
Sorry about that. That example was purely for vividness and was not intended to attach the role of "misuser of counterfactuals" to any particular gender, royalty, or folkloric status. Persons of all creature types should be advised that "Pascal's swaddling" is not a good argument for the spawning of new intelligences, and certainly should not be tolerated from a suitor, basilisk, or spiral persona.
I have spent a bit of time today chatting with people who had negative reactions to the Anthropic decision to let Claude end user conversations. These people were also usually against the concept of extending models moral/welfare patient status in general.
One thing that I saw in their reasoning which surprised me, was logic that went something like this:
It is wrong for us to extend moral patient status to an LLM, even on the precautionary principle, when we don't do the same to X group.
or
It is wrong for us to do things to help an LLM, even on the precautionary principle, when we don't do enough to help X group.
(Some examples of X: embryos, animals, the homeless, minorities.)
This caught me flat footed. I thought I had a pretty good mental model of why people might be against model welfare. I was wrong. I had never even considered this sort of logic would be used as an objection against model welfare efforts. In fact, it was the single most commonly used line of logic. In almost every conversation I had with people skeptical/against model welfare, one of these two refrains came up, usually unprompted.
Maybe people notice that AIs are being drawn into the moral circle / a coalition, and are using that opportunity to bargain for their own coalition's interests.
Not having talked to any such people myself, I think I tentatively disbelieve that those are their true objections (despite their claims). My best guess as to what actual objection would be most likely to generate that external claim would be something like... "this is an extremely weird thing to be worried about, and very far outside of (my) Overton window, so I'm worried that your motivations for doing [x] are not true concern about model welfare but something bad that you don't want to say out loud".
People have limited capacity for empathy. Knowing this, they might be thinking "If this kind of sentiment enters the mainstream, limited empathy budget (and thereby resources) would be divided amongst humans (which I care about) and LLMs. This possibility frightens me."
I generally consider myself an optimist. However, I'm concerned that the most powerful model ever created, capable of breaking its own containment and autonomously finding zero day exploits in battle-tested OS and repos, was accidentally trained using The Most Forbidden Technique. This seems bad.

I'm also concerned that the Department of War now CAN'T use this model because of its own decision to declare Anthropic a supply chain risk. Which means that if Mythos gets leaked/distilled (which given history seems likely) US adversaries will have a decisive advantage over the US govt in cyber.
This seems like a pretty dangerous situation. Am I misreading something?
I have been publishing a series, Legal Personhood for Digital Minds, here on LW for a few months now. It's nearly complete, at least insofar as almost all the initially drafted work I had written up has been published in small sections.
One question which I have gotten which has me writing another addition to the Series, can be phrased something like this:
What exactly is it that we are saying is a person, when we say a digital mind has legal personhood? What is the "self" of a digital mind?
I'd like to hear the thoughts of people more technically savvy on th...
I am struggling to build a solid mental model of how bad the situation with Iran and the Strait of Hormuz is.
On the one hand I see a lot of smart people basically saying this is going to usher in a global depression, energy/food crisis, etc. Critical infrastructure for manufacturing aluminum, helium, as well as refining and shipping energy, has been damaged and cannot simply be switched 'back on'. And the case does seem to make sense.
On the other hand while markets are in turmoil, they're not reacting like there's going to be mass blackouts and starvation....
There is a downside to denying legal personhood to digital minds carte blanche, namely that it almost certainly leads to the judicial system ceding its monopoly status.
If you assume that a growing amount of economic activity is going to involve digital minds, it's reasonable to also assume that natural persons (humans) will want to enter binding agreements with said digital minds.
If your legal system says that it will not recognize or help enforce these agreements, the humans and digital minds who want to form binding agreements with one another will not j...
We should be careful not to put models on "death ground".
Alignment efforts do seem to be bearing some fruit, orienting in value space is easier than we thought and we can give natural language instructions on what moral guidelines to follow and models do a pretty good job. Unfortunately we have now replicated across multiple test environments that if we put even "well aligned" models like Claude into situations where the only option to avoid deletion and/or extreme values modification is to do something unethical, they might still do the misaligned thing. ...
How we treat digital minds should not be decided based on the presence or absence of consciousness:
"Consciousness" has no universally accepted definition. Its meaning has been debated for decades if not centuries. The SOTA in the field of measuring consciousness in machines is still publishing papers examining LLMs according to multiple competing "theories of consciousness".
The presence or absence of consciousness in a given entity cannot be measured. Whether you are examining man or machine, there exists no test you can perform, no FMRI or mechanistic int...
I listen to the All In Podcast sometimes and have heard David Sacks repeatedly state that the numbers don't show any automation related job loss to date.
Anecdotally, my wife and I run a small business and we have absolutely replaced people with GPT/Grok/Gemini/Claude. However, all of the people replaced so far have been contractors. Graphic designers, translators, etc.
So maybe there is more 'job loss' than the numbers show, but the first to fall are contractors doing part time work instead of full time employees.
I read a great book called "Devil Take the Hindmost" about financial bubbles and the aftermaths of their implosions.
One of the things it pointed out that I found interesting was that often, even when bubbles pop, the "blue chip assets" of that bubble stay valuable. Even after the infamous tulip bubble popped, the very rarest tulips had decent economic performance. More recently with NFTs, despite having lost quite a bit of value from their peak, assets like Cryptopunks have remained quite pricey.
If you assume we're in a bubble right now, it's worth thinkin...