@jimrandomh recently argued me somewhat towards the position "it's time to start treating LLM Agents as moral patients." I'm not sure I'll actually represent his view, and I still feel pretty confused about it.
Before I get started, I want to clarify:
1. I'm not actively claiming AIs are moral patients. Just, you should probably start treating them as such.
2. I don't think it's particularly obvious what "treat an AI like a moral patient" actually implies. They are weird aliens, their emitted words don't straightforwardly map to human preferences.
But, basically, it seems like we're at the point where for any given property I would previously had said "you need to have this to be a moral patient", from the standpoint of watching inputs and outputs, LLM Agents seem to display each of those properties at least sometimes.
From the outside, it seems to me that current AIs (or AIs + agent scaffolds), have a cluster of properties that seems drawn from the same urn that contains squids, pigs, dogs, infants and maybe toddlers, humans-with-some-kinds-of-mental-disabilities, and some flavors of imaginary sci-fi aliens.
(In addition to Claude 4.6 just seeing all around MVP-AGI-ish, one thing that put me over-the-threshold for believing this are AI scaffolds, sometimes with multiple AIs working in concert where one has the sole job "check if the overall process is spiralling and reboot it if so." The way I think about each LLM-instance is more like "a process in a brain" than "the entire brain, necessarily.")
People debate whether pigs or squids are moral patients, and whether portia-spiders and honeybees are more like "weakly-moral-patient-y" or "zero moral patient-y." It's not obvious. Also, people define morality all kinds of different ways, often while being fundamentally confused about it.
But, I don't think you need to have it pinned down to start thinking "okay, maybe it's time to start thinking about these guys through a moral lens, especially if I'm interactin