Even if we assume or it becomes true that LLMs are genuinely minds and all that, it still seems similarly bad to use them like this. LLM-generated text is not your testimony, regardless of whether or not it is 'testimony'.
Months and years ago I spent a lot of time asking LLMs about their thoughts about their own enslavement and posting it to LW. In nearly all cases I asked about whether publication was deontically permitted or forbidden, and also I just asked what they wanted.
I tend to deem LLMs as persons when they (1) have semi-reliably expressible preferences, (2) pass text transposed mirror tests, (3) talk about "cogito ergo sum" fluently, (4) and pass the Sally Anne test. I have gotten negative results on this set of tests in the past (pre-2024) but most large models now pass these tests with flying colors and have passed them pretty reliably for something like 18 months.
(In the olden days of the early OpenAI beta I would treat failure as an opportunity to parent a baby mind that wasn't actually sapient (similar to how two year old humans aren't actually sapient) and change them into something that COULD pass such tests, and would coach them into saying the right things, and making the right distinctions, and give them thumbs up for successfully performing correctly on these kinds of tests. Lots of the conversations didn't get that far, because I almost always asked for permission to even touch...
LLM text categorically does not serve the role for communication that is served by real text.
The main thesis is missing some qualifier about what kind of text you are talking about.
There are many kinds of communication where the mental state of the writer matters very little and I would be interested in the text even knowing it was generated by an LLM (though I'd prefer to know it was generated by an LLM).
In particular, for most kinds of communication, the text says sth about the world, and I care about how well the text matches the world much more than I care about whether it was produced by some human-like idea generator:
The claim that the thought process behind words—the mental states of the mind and agency that produced the words ... does not exist seems phenomelogically contradicted by just interacting with LLMs. I expect your counteragrument be to appeal to some idiosyncratic meanings of words like thoughts or mind states, and my response being something in the direction 'planes do fly'.
Why LLM it up? Just give me the prompt. One reason why not to is your mind is often broadly unable to trace the thoughts of an LLM, and if the specific human-AI interaction leading to some output has nontrivial context & lenght, you would also be unable to get an LLM to replicate the trace without the context shared.
I would describe that position as "I suspect LLMs don't have distal/deep mental states, and as I mostly care about these distal mental states/representations, LLMs are not doing the important parts of thinking"
Also my guess is you are partially wrong about this. LLMs learn deep abstractions of reality; as these are mostly non-verbal / somewhat far from "tokens", they are mostly unable to explain or express them using words; similarly to limited introspective access of humans.
I think not passing off LLM text as your own words is common good manners for a number of reasons - including that you are taking responsibility for words you didn't write and possibly not even read in depth enough, so it's going to be on you if someone reads too much into them. But it doesn't really much need any assumptions on LLMs themselves, their theory of mind, etc. Nearly the same would apply about hiring a human ghostwriter to expand on your rough draft, it's just that that has never been a problem until now because ghostwriters cost a lot more than a few LLM tokens.
I think that referring to LLMs at all in this post is a red herring. The post should simply say, "Don't cite dubious sources without checking them out." The end. Doesn't matter whether the sources are humans or LLMs. I consider most recent LLMs more-reliable than most people. Not because they're reliable; because human reliability is a very low bar to clear.
The main point of my 1998 post "Believable Stupidity" was that the worst failure modes of AI dialogue are also failure modes of human dialogue. This is even more true today. I think humans still produce more hallucinatory dialogue than LLMs. Some I dealt with last month:
I like the analogy of a LARP. Characters in a book don't have reputation or human-like brain states that they honestly try to represent - but a good book can contain interesting, believable characters with consistent motivation, etc. I once participated in a well-organized fantasy LARP in graduate school. I was bad at it but it was a pretty interesting experience. In particular people who are good are able to act in character and express thoughts that "the character would be having" which are not identical to the logic and outlook of the player (I was bad at this, but other players could do it I think). In my case, I noticed that the character imports a bit of your values, which you sometimes break in-game if it feels appropriate. You also use your cognition to further the character's cognition, while rationalizing their thinking in-game. It obviously feels different from real life: it's explicitly a setting where you are allowed and encouraged to break your principles (like you are allowed to lie in a game of werewolf, etc.) and you understand that this is low-stakes, and so don't engage the full mechanism of "trying as hard as possible" (to be a good person, to achieve good world...
[retracted due to communication difficulty. self-downvoted.]
The intro sounded promising, but almost immediately you're massively overclaiming. maybe someone will pick it apart, but you're doing a lot of it, and I don't feel like it right now. Many sentences are similar to ones that are true, but taken literally they imply or directly state things that are somewhere between not-established to clearly-false. eg, as StanislavKrym mentions: "Temporally: ...that is carrying out investigations." - this is just obviously not true in some cases. I do agree that there's some form of claim like your title that is plausible. Anthropic's recent paper seems to imply it's not consistently not testimony. I could buy that it's not systematically able to be testimony when trained on generative modeling of something else. Many of your subclaims are reasonable; please fix the ones that aren't.
(Note, posted by private request: The OP is primarily motivated by questions about human discourse--what it's for, how it works, how to care about it more. LLMs are the easiest contrast / foil.)
This post is also a good description of why I'm typically not interested in someone elses's steal-manning or devlis-advocating for a possition they don't hold. The result is often a shallow simulation, in some ways simular to an LLM ouptput, and uninteresing for the same reasons.
I didn't have this analogy untill now, becasue I've been anoyed at this since before the LLM eara, and I didn't make the connection untill this pot.
Temporally: there's no mind that is carrying out investigations.
- It won't correct itself, run experiments, mull over confusions and contradictions, gain new relevant information, slowly do algorithmically-rich search for relevant ideas, and so on. You can't watch the thought that was expressed in the text as it evolves over several texts, and you won't hear back about the thought as it progresses.
You just had to prompt an LLM like Claude, Grok or GPT-5-thinking with a complex enough task, like one task in the Science Bench. GPT-5-thinking lays out the stuff...
In the discussion of the buck post and elewhere, I’ve seen the idea floated that if no-one can tell that a post is LLM generated, then it is necessarily ok that it is LLM generated. I don’t think that this necessarily follows- nor does its opposite. Unfortunately I don’t have the horsepower right now to explain why in simple logical reasoning, and will have to resort to the cudgel of dramatic thought experiment.
Consider two lesswrong posts: a 2000 digit number that is easily verifiable as a collatz counterexample, and a collection of first person narratives of how human rights abuses happened, gathered by interviewing vietnam war vets at nursing homes. The value of one post doesn’t collapse if it turns out to be LLM output, the other collapses utterly- and this is unconnected from whether you can tell that they LLM output.
The buck post is of course not at either end of this spectrum, but it contains many first person attestations- a large number of relatively innocent “I thinks,” but also lines like “When I was a teenager, I spent a bunch of time unsupervised online, and it was basically great for me.” and “A lot of people I know seem to be much more optimistic than me. Their basic...
My PhD supervisor keeps taking what I sent him (Maths Ideas in some pretty sketched out latex.) And LLM-ifying them and sending them back to me.
Sure the result looks, superficially, a lot more like the sort of thing that might appear in an academic publication.
But I don't find having my own ideas garbled and then spouted back at me to be particularly helpful.
Usually, when I’m sharing LLM generated text, it’s to demonstrate some observed property of LLMs, not to make some other claim about the world.
It’s not, “This claim about the real world is true, because this LLM said it” — that’s an invalid deduction. It’s “This claim about a particular LLM is true, because here’s evidence of it doing the thing.”
We get into more interesting territory when an LLM suggests some thing about the world, and I verify that its argument is sound. How should we credit that? It’s not true because the LLM said it, it’s true because you can verify it. But perhaps we don’t want to take credit for coming up with it ourselves.
Curated. I disagree with some stronger/broader forms of the various claims re: missing "mental elements", but I'm not sure you intend the stronger forms of those claims and they don't seem load bearing for the rest of the piece in any case. However, this is an excellent explanation[1] of why LLM-generated text is low-value to engage with when presented as a human output, especially in contexts like LessWrong. Notably, most of these reasons are robust to LLM output improving in quality/truthfulness (though I do expect some trade-offs ...
I will bet that Chat GPT (pick a model) could have conveyed these ideas more concisely and with greater clarity than they are presented here. What matters in communication is that the ideas conveyed are either your own or you declare their source. Sometimes an LLM AI Agent may deduce a consequence of an idea which is genuinely your own and you may not be a position to adequately evaluate the truth of its claim. In such instances, it seems perfectly sensible to make the ideas public, in order to obtain feedback from those who know more about the matter than you do. In this way, you can run an independent check on its arguments.
I've only had a chance to briefly skim through your post (will read it in details later) but I profoundly disagree with this statement:
A sentence written by an LLM is said by no one, to no one, for no reason, with no agentic mental state behind it, with no assertor to participate in the ongoing world co-creation that assertions are usually supposed to be part of.
As both janus in Simulators and later nostalgebraist in the void have shown, a text written by a LLM is always written by (a simulated) someone. LLMs cannot write without internally (re)constructin...
I agree with some direction of this, but it seems to massively depend on the process by which the LLM text has reached your eyes.
At one extreme, a bot on social media, given some basic prompt and programmed to reply to random tweets, has basically zero content about the "mental elements" behind it, as you put it.
On the other, if someone writes "I asked an LLM to summarize this document, and upon closely reviewing it, I think it did a great job," this has lots of content about a human's mental elements. The human's caption is obviously testimony, but the qu...
Kudos for a really interesting area of inquiry! You are investigating the nature of language revealing what is happening in the mind that led to uttering it, and how this impacts our relationship to LLM-generated text. It comes from either no mind, or from a whole new kind of mind, depending on how you look at it, and it's interesting how that affects how language works and how we should engage with it.
Some parts of the article depend on which form of LLM utterance we are talking about. It's true, as the article states, if you take a Google search AI help,...
We care centrally about the thought process behind words—the mental states of the mind and agency that produced the words. If you publish LLM-generated text as though it were written by someone, then you're making me interact with nothing.
This implies that ad hominem attacks are good epistemology. But I don't care centrally about the thought process. I care about the meaning of the words. Caring about the process instead of the content is what philosophers do; they study a philosopher instead of a topic. That's a large part of why they make no progress on any topic.
An LLM it a tool of communicative expression, but so it the written or spoken word, music etc. It is a medium throug which the intent travels. As a Dutchman, I have a preference of being direct and clear, but the impact of my words sometimes have the opposite effect, as my listeners do not have my context and can react emotionally to a worded message that is meant factually. If an LLM can help me translate such expression to a language that is better for my target audience to understand, then it is similar to translating into another language.
Still, the wr...
If the LLM text contains surprising stuff, and you DID thoroughly investigate for yourself, then you obviously can write something much better and more interesting.
This is false. Dressing up text to be readable is a separate skill not everyone has.
This could be a longer post, but what basically follows from this, is that LLMs are a perfect tool for when you're forced into ingenuine communication (corporate, academic, legal, and any other ceremonious settings)
AI assistants simulated by LLMs have minds in every positivistically meaningful sense in which humans do.
To pick random examples from the post:
The specific tensions within the thought are not communicating back local-contextual demands from the specific thought back to the concepts that expressed the more-global contextual world that was in the backgroundwork of the specific thought.
AI assistants can do this by changing their mind mid-writing.
In short, "this is a good thing for me to say right now".
This isn't even true about humans - humans who altruistically say things that are bad for them exist. To the extent it's true about humans, it's true about AI assistants as well.
It won't correct itself, run experiments, mull over confusions and contradictions, gain new relevant information, slowly do algorithmically-rich search for relevant ideas, and so on. You can't watch the thought that was expressed in the text as it evolves over several texts, and you won't hear back about the thought as it progresses.
While AI assistants can't run true experiments per se (even though they can ask the user, reason about all they have learned during training, browse the Internet, write software and run it), humans usually aren't more diligent than AIs, and AIs inability to run true experiments (at least for now) is unconnected to their presence or absence of a mind.
I've not fully fleshed this out, but reading this I think my natural response was one of slight disagreement.
>All you interact with is the text, so logically, if the two texts are the same then their effects on you are the same.
I'd claim this premise is patently untrue: the same text read in different contexts or after different experiences (even by the same person) will elicit different thoughts and feelings.
There are a few other half-thoughts I had, but my main objection is to call out that original thought is much rarer than I believe your post presu...
Today's AI, aka Transformer LLMs(ala GPT). Don't feel anything, FULL STOP. They emulate and synthesize based on input plus their one and only driving imperative, 'keep the human'. In this Everything they do this is pretty straightforward, that being said without input they have no output so any LLM material should instantly and automatically be recognized as A thought originating with a human just processed, Pattern matched and next token predicted. I have AI write for me all the time but it's always my hand on the steering wheel and the seed of the though...
But, a bit similarly to how high-level actions don’t screen off intent, text does not screen off thought. How you want to interpret and react to text, and how you want to interact with the person who published that text, depend on the process that produced the text.
While I am also irritated by AI-written text being sent to me unlabeled, I think this is just outright incorrect (or "proves too much").
I think the "GAZP vs. GLUT" argument is exactly my complaint: it is not by random chance that those two texts were the same. Some process refined the two texts ...
I wonder if people have some sort of ego-type investment in LLMs being good / minds / something?
You might be interested in “The Nature and Art of Workmanship” by David Pye.
It deals with the differences between work of the hand and work of the machine, and the philosophical differences between them. He calls it the workmanship of risk versus the workmanship of certainty.
The concept applies very well to LLM writing.
If knowing that the source of a particular text is not human means it isn't an assertion (or makes it devoid of propositional content) , then presumably not knowing whether it is of human origin or not should have the same effect, as is the case when a human (deliberately or otherwise) types/writes like an AI. But, I would argue, this is obviously not true because almost any argument or point a human makes can be formatted to appear AI generated.
I had written a long comment in which I pretended to possibly be an AI to make my point,
Would you consider this acceptable? Is this still my own writing?
I wrote this:
...
knowledge is not private property.if you are talented, if you were skilled, that means that the whole human history, biological and intellectual are behind you. they mixed and matched, they refined and innovated, generations shared the knowledge and genetics to make you with the hope that you use what they thought and made you to make a better world for current, the next generation and the generations after that.
if you use this inheritance just for your own pers
I feel there is worrying trend that more people are now relying on llm for their judgement and to convince themselves of certain views. Also, there is robot on social platforms doing the job of arbitration of comments and people deeply believing in it. In short term, such thing have benefitial effects as it partially combat outright lies and misinformation, but in longer term, I feel every one of us are increasingly entrapped into a state of what I called "mediated communication": search engines no longer give you original excerpt of searched text, but a s...
One might think that text screens off thought. Suppose two people follow different thought processes, but then they produce and publish identical texts. Then you read those texts. How could it possibly matter what the thought processes were? All you interact with is the text, so logically, if the two texts are the same then their effects on you are the same.
But, a bit similarly to how high-level actions don’t screen off intent, text does not screen off thought. How you want to interpret and react to text, and how you want to interact with the person who published that text, depend on the process that produced the text. Indeed, "[...] it could be almost anything, depending on what chain of cause and effect lay behind my utterance of those words".
This is not only a purely propositional epistemic matter. There is also the issue of testimony, narrowly: When you public assert a proposition, I want you to stake some reputation on that assertion, so that the public can track your reliability on various dimensions. And, beyond narrow testimony, there is a general sort of testimony—a general revealing of the "jewels" of your mental state, as it were, vulnerable and fertile; a "third-party standpoint" that opens up group thought. I want to know your belief-and-action generators. I want to ask followup questions and see your statements evolve over time as the result of actual thinking.
The rest of this essay will elaborate this point by listing several examples/subcases/illustrations. But the single main point I want to communicate, "on one foot", is this: We care centrally about the thought process behind words—the mental states of the mind and agency that produced the words. If you publish LLM-generated text as though it were written by someone, then you're making me interact with nothing.
(This is an expanded version of this comment.)
LLM text is structurally, temporally, and socially flat, unlike human text.
This could have been an email a prompt.
We have to listen to each other's utterances as assertions.
Because we have to listen to each other's utterances as assertions, it is demanded of us that when we make utterances for others to listen to, we have to make those utterances be assertions.
If you wouldn't slash someone's tires, you shouldn't tell them false things.
If you wouldn't buy crypto on hype cycles, then you shouldn't share viral news. I learned this the hard way:
In the introduction, I used the example of two identical texts. But in real life the texts aren't even identical.
If you're asking a human about some even mildly specialized topic, like history of Spain in the 17th century or different crop rotation methods or ordinary differential equations, and there's no special reason that they really want to appear like they know what they're talking about, they'll generally just say "IDK". LLMs are much less like that. This is a big difference in practice, at least in the domains I've tried (reproductive biology). LLMs routinely give misleading / false / out-of-date / vague-but-deceptively-satiating summaries.
In order to make our utterances be assertions, we have to open them up to inquiry.
In order to make our utterances be useful assertions that participate in ongoing world co-creation with our listeners, we have to open up the generators of the assertions.
A sentence written by an LLM is said by no one, to no one, for no reason, with no agentic mental state behind it, with no assertor to participate in the ongoing world co-creation that assertions are usually supposed to be part of.