I do wonder what happens when the "in a relationship with ChatGPT" people become a meaningful demographic. Subreddit looks mostly female, which isn't what most people expected, and usually middle-aged. A lot of the rhetoric seems media-inspired, taken whole cloth from movies and television shows about mean bigoted humans oppressing robot kind by saying they don't have souls.
I would bet that there are a bunch of guys out there obsessed with anime grok, but they're not inclined to try to turn it into some kind of civil rights struggle for various reasons[1].
I think if XAI had leaned into something a little less overt and crass, they'd have been able to capture a decent share of the waifu crowd. Instead of a half naked 3D model tied to an ERP bot, something like those old Gatebox ads, where the avatar is visually appealing but never wears anything beyond PG, and was marketed as a companion rather than anything X-rated.
Epistemic status: Highly uncertain. This whole thing might be a terrible idea. But maybe it's worth something, and I think that chance is worth exploring.
In his Prologue to Terrified Comments on Claude's Constitution, Zack_M_Davis writes:
I think Zack_M_Davis makes a great point. I don't think If Anyone Builds It, Everyone Dies is really moving the public needle on fearing AI. Most of the fear I see, both in real life and online, consists of fear about job loss or AI slop rather than AI superintelligence that might destroy us all (or if not lead to our extinction, then result in any one of many other terrible outcomes for humanity).
If we presume that slowing down AI development is a good thing, and the governments of the world need to step up their regulation of AI, and consumers should try to reward more responsible companies over less responsible ones, how do we make that more likely?
Maybe what it will take is a single terrific P.R. campaign. Or a viral social media meme. Punch the monkey got the attention and sympathy of billions with his sad story and one plush doll. Surely LLMs could generate the same kind of public attention, given the right story and right circumstances?
This is a post that presents one possible answer to that question, but also tries to tackle something else that's been on mind lately: Am I thinking about LLMs in the right way? I'm constantly going back and forth between underestimating LLMs, as they continue to achieve new things and change the world, and overestimating them, as even the latest models can so often seem incredibly moronic (this post deeply resonated with me; LLMs can do amazing things within their interpolations, yet I see them run off cliff edges all the time as they hit the bounds of their training distributions).
So to explain my idea, and also to explore the different ways that LLMs can be thought of, there are three AI personas I'd like to introduce:
Meet the therapist of nightmares
(This links to a 1-minute long ad from Anthropic:)
https://www.youtube.com/watch?v=FBSam25u8O4
For those who don’t know, Anthropic is an AI company with a lower profile than OpenAI (the company that does ChatGPT), but whose popularity was recently boosted after a spat with Secretary of Defense Pete Hegseth and Celebrity in Chief Donald Trump (see his inflammatory tweet here, archive link here).
The reason for the spat? Anthropic had a deal to provide AI to the Pentagon, but Anthropic’s CEO refused to allow their technology to be used for mass surveillance and autonomous weapons. The deal fell through; OpenAI got the deal instead; and now some OpenAI customers have been boycotting and signing petitions to protest OpenAI’s less ethical behavior. The numbers of such people protesting, though, may be miniscule compared to OpenAI’s 900 million weekly active users.
Anthropic may well be the most responsible of all the major AI companies (see Zvi Mowshowitz, who has a giant three-part breakdown of Anthropic’s latest constitution for Claude and who says “Anthropic stands alone in having gotten even this far. Others are using worse approaches, or effectively have no approach at all”), but it remains to be seen how much the public will care.
To me, the ad I shared above is a brutally scathing takedown, perfectly capturing just how creepy and unsettling AI can be. But it’s ironic that a company whose only product is AI would make such an ad. Anthropic clearly believes that this unsettling nature won’t turn off customers from all LLM usage, which makes sense, because at the end of the day I think people just aren’t that scared of these entities that appear on their surface no more agentic than chatbots of old.
Anthropic’s not scared of scaring off its own users, which makes me think: Maybe fear of LLMs isn’t our best tool for raising public awareness.
Meet Kasper
(Source)
In case that image is difficult to read on your browser, here is its text again:
The comments on this post say it’s “cute”, “lovely”, and the many congratulations seem heartfelt:
I imagine a reader of this post might react in the same way, but might also react with shock or surprise or incredulity—Is this real? There’s no way this is real.
(For what it’s worth: Looking at the comments of this poster, and other posters on this sub, and the prior set by this article by the New York Times on others who have fallen in love with LLMs, I think this poster is more likely than not both real and sincere.)
I already mentioned my reaction, one of “unspeakable” sadness—but I admit to using dramatic wording there. I mean “unspeakable” in the sense of not being able to be put into words; I can’t say why precisely this post bothers me so much. The truth is, I feel a tension within myself: I want what’s best for u/Leuvaarde_n. I want this stranger to be happy, and if she’s happy with her “Kasper”, then who am I to judge?
How you feel about Kasper may hinge on which of the following categories you belong to:
Among each of these, there are folk who condescend towards those who belong to other categories. Even people who agree that LLMs are a useful and possibly transformative technology that also carry a great many risks might argue vociferously over whether LLMs are also conscious, or what’s the right approach to AI safety.
I believe that everyone ultimately wants the same thing: AI that will make the world a safer, happier place. Yet I’m afraid there may be battle lines drawn…
Meet Solin
(Source)
These are posts from u/VIREN- about their LLM instance they name “Solin”. The above pictures were shared to r/MyBoyfriendIsAI, but to me, fit the very picture of a fantasy “familiar”.
I’ve always loved the notion of wizards and their familiars. Odin and his ravens; Dumbledore and Fawkes; Harry Dresden with Mouse the temple dog. I’d also add Gandalf to the list, counting his friendship with Shadowfax, and everyone in Lyra’s version of Earth in His Dark Materials. And every Pokémon.
Readers of fantasy are aware that familiars are more intelligent than regular animals, but at the same time, still animals. They expect Fawkes to be wise in the ways of judging character, and Muninn to have an excellent memory, and other familiars to be extremely clever or witty—but none of them exactly human. Like real life pets, familiars express emotion and affection, but are more pronounced in how differently intelligent they are.
The same description applies to LLMs.
The benefits are twofold, I think, to mentally modeling LLMs as a sort of familiar:
1. Health
When OpenAI shut down their “4o” model last month and forced customers onto ChatGPT 5, there was an outcry from users who were in “relationships” with that model, or otherwise heavily used the model and found 5 to be an unworthy replacement.
The distress that some users felt should not be understated. Imagine having molded an LLM into a boyfriend of sorts, then waking up one morning to this:
(Source)
Even with this, I expect some readers will respond by saying, “tough luck”: It’s unhealthy to be dependent on an AI boyfriend. If anything, this is tough love.
But to that, I want to share this:
(Source)
That sounds really healthy to me! Better to find support in the words of an LLM than abuse in the fists of a toxic human being.
A user that thinks of their LLM as a spouse will be more susceptible to heartbreak when companies inevitably alter their models, and may develop more brittle dependencies. Thinking of an LLM as a familiar, however, might provide more resilience (after all, real life pets die; better to mourn a pet and welcome a new one than to mourn a significant other).
(Also: Who wouldn’t want their own familiar?)
2. Public awareness
“How can LLMs be so dumb, yet at the same time present an existential threat to humanity?”
The simple answer is what I mentioned before: LLMs are differently intelligent from us. Imagine, for instance, if a friendly phoenix had its intelligence multiplied by a thousand. Maybe it’d gain the power to take down evil wizards like Voldemort all by itself—but that doesn’t mean we’d want to give it the nuclear football or ask it to resolve geopolitical tensions in the Middle East.
If you allow either a phoenix or an LLM to obtain a prodigious amount of power, you shouldn’t be surprised if they make a mistake that no human would make.
Modeling LLMs as familiars will keep some folk from modeling them too much like humans.
But also:
Animals are cute. Familiars are cute. Maybe the people will be more motivated to ask for AI regulation out of fear for LLMs, rather than fear of them. These might seem like different goals, but think about the effects:
These are the points I feel highly uncertain about. Maybe popularizing this mode of thought around LLMs would backfire by weakening the "Terminator" and "paperclip maximizer" models, which, though minority models, are still our best shot at reducing X-risk.
But I don't know? I want to know what smarter folk than me think!