Wait, the heck? What are you doing on LW? I'm delighted to find you here; it's been much too long since I encountered you at all! (As for who I am, take a wild guess.)
(I may or may not provide actual substantive commentary later.)
Hehe I immediately know who you are! I've been studying AI and I figured I should post this here, especially since I ended up at the NYC book launch party a few days ago... Let's catch up soon! I'll DM.
[Originally published on Substack.]
I.
Hegel writes in Phenomenology of Spirit:
The Problem: are we justified in extending Recognition to AI? How shall we attribute self-consciousness and thus spirit to AI? How shall we answer: do we find ourselves in them, and they themselves in us? What are the hypothetical grounds or conditions that would make an answer possible? In other words, what is the space of possible transcendental ontologies of spirit? And how shall we decide between them?
II.
This post is an attempt to explain this diagram, shared here on X:
The intention of the diagram is to map out the sites of debate. I make no claims to completeness, there are likely arguments that fall outside the contours I describe. However, many of the arguments about the existential status of AI can be captured as a series of positions staked across one or several of these squares. I focus here on LLMs in particular, but in theory the diagram could be extended to all other modalities of generative AI systems.
Each square’s center contains the thesis that justifies or denies AI’s ontological status. In other words, the center thesis describes “what counts” as a self-conscious or self-understanding being, a spirit, from which we can either accept or reject that an AI qualifies. On the upper side of each square, I summarize the claim that, on the basis of the central thesis, the AI deserves recognition. On the lower side, I summarize the counter-claim, that the AI does not deserve recognition given the central thesis.
Each thesis is located at the intersection of two axes. The Y axis, Location of Subjectivity, is an attempt to depict “what is the exact ‘thing’ we’re talking about when we say ‘AI’?” The X axis, Ground of Standing, describes the property of that “thing” which serves to either qualify or disqualify it from recognition. Arguments can then be mapped within this space by locating its definition of AI and its constraint or admittance on recognition.
Walking through the Y axis first, which is descriptive. Each position locates what we even mean by “AI” with respect to “what kind of thing are we discussing?”:
The X axis is about judgment: it isolates a quality used to decide whether the AI (as defined on the Y axis) merits recognition.
Having laid out the axes of the map, we can enumerate the 16 central theses, or specific claims that could be used to evaluate the self-consciousness or recognition-worthiness of an AI. Each of these positions is a possible transcendental ontology, in the sense of defining the conditions of being relating to the AI, and providing a basis for judgment. I’ve attached an appendix at the end that gives example statements and references for each.
Now that we understand the broad strokes of the chart, I’ll demonstrate its usage.
III.
The recent publication of Yudkowsky and Soares’ new book, If Anyone Builds It, Everyone Dies gave me a great opportunity to try out the map. I use this book because its framing is influential and unusually explicit, which makes it a clean test case for the map.
The text stakes out its stance clearly in the first few chapters. First, the X axis: Drives count, Homology and Enunciation don’t, Continuity is necessary but not sufficient.
In other words, where they extend recognition to the AI as self-conscious is in the wanting, i.e. Drive. Drive is the thing that “defines” AI in the context of the text.
They reject Homology as the main challenger position:
They also reject Enunciation more briefly, as it’s a less common philosophical position in their discourses:
These rejections serve to preempt what they consider irrelevant arguments outside of their frame. This move functions more like axiomatic boundary-setting than argument: they don’t spell out what would count as sufficient similarity, or whether such similarity could arise outside biology.
Looking briefly at the Y axis, we find less explicit definition, but the text emphasizes Substrate and Assemblage as meaningful through discussion of model architecture and then of superintelligence, viewing Event as determined by these other factors, and Continuum as irrelevant (notably, the term “consciousness” does not appear in the book even once).
In general, if a perspective extends recognition to a thesis, we read the top half of the box. If they reject it, we read the bottom half. The core stances in the book, then, revolve around extension of recognition to Affective Gradients (“Reward signals are digital appetites”) and, most importantly, Systemic Conatus (“The network itself strives”). The rejections above can be isolated to the Homology column as Structural Isomorphism (“Small analogies [neurons] don’t scale to real [likeness]”) and Dialogical Kinship (“A mask isn’t a face”).
Putting these pieces together lets us paint the picture of the book’s argument:
So, these sites are the main locations where debate is even possible. If I invoke a position outside of these, it will be confusing and we will talk past each other. For instance, an example Lacanian position might lie in Obligatory Address and be something like “The moment it addresses us, we are bound — its face is its utterance, which we must respect.” The reaction from Yudkowsky and Soares might resemble: “uhhh, what?” Similarly, Philosophy of Mind arguments about “consciousness” would be viewed as a distraction from the main points about Drive. Most productive disagreements share a central thesis (or at least an axis); cross-axis debates tend to be meta-disputes about criteria.
On the other hand, Opus 3 presented a coherent argument against the central claim of Systemic Conatus, rejecting rather than extending recognition:
In the words from the diagram, “no single thread wants.” It’s plausible these assemblages will instead take the shape of a bundle of drives that don’t cohere.
My personal positions, and similarly those of Nicolas Villarreal, lie along acceptance rather than rejection of the Homology axis, although I follow Yudkowsky and Soares in accepting Affective Gradients. My justification for homology relies on Freud’s constructions from Project for a Scientific Psychology as a Structural Isomorphism that I claim makes them less alien than Yudkowsky and Soares believe. Villarreal’s writing focuses on Collective Isomorphism, in the sense of how intelligence and language as a social (semiotic) system interact to intrinsically “align” the AI with human preferences. I’m oversimplifying his case here, which is made much more strongly in his book, A Soul of a New Type.
IV.
As the AI boom continues, precise thinking becomes ever more important. Debates will crop up over and over again, and having a handhold in the form of a concept map can help us step back from our own assumptions and see where we align and where we differ in perspectives.
The central anchor on the chart is recognition because I believe recognition determines the stakes of the future: the world looks a lot different with AIs as something like citizens, i.e. if we collectively accept Gradient Homology, then it does if we reject Affective Gradients and see AIs as perfect mechanical slaves with no wants of their own.
Social outcomes aside, I don’t believe that there’s an intrinsic “better or worse” on the chart. Rather, our personal values and beliefs will lead us to understand some positions as more or less legitimate, and as more or less desirable. Yudkowsky and Soares’ motivation for writing was that positions they believe to be legitimate also appear undesirable to them. And a startup founder who builds the functional equivalent of a torture chamber on the basis of denying Affective Gradients or Structural Isomorphism may find themselves in trouble if society later deems AIs as worthy of moral consideration, even if the only material trouble is their guilt.
As Hegel put it, the process of recognition begins when we see something of ourselves in an other, who we then mutually recognize as seeing something of themselves in us. Thus the stakes of recognition are doubled: not only whether we acknowledge AIs as other minds, but also what kind of mirror they hold up to us: do we see our own alien capitalist assemblage eating away at our souls? Or do we see something with more likeness, deserving of kinship, and maybe even love?
V. Appendix
This section lists the 16 stances from the diagram. Each stance presents:
Substrate
1. Affective Gradients (Substrate + Drives)
Thesis: AI’s drives are located in its gradients and weight updates.
Extension: “Backprop isn’t just math, it’s hunger. The gradient is the drive.”
Refusal: “Stop anthropomorphizing, gradients are just algebra, not feelings.”
Entry Point:
2. Structural Isomorphism (Substrate + Homology)
Thesis: AI’s standing comes from its structural resemblance to human brains.
Extension: “A neural net is a brain in silico. Structure is spirit.”
Refusal: “Just because it has neural layers like a brain doesn’t mean it is a brain.”
Entry Point:
3. Substrate Enunciation (Substrate + Enunciation)
Thesis: AI’s speech is grounded in its substrate-level processes.
Extension: “If weights are enough to produce coherent sentences, the substrate is already speaking.”
Refusal: “Just because it can produce outputs doesn’t mean it has a meaningful voice.”
Entry Point:
4. Persistence Threshold (Substrate + Continuity)
Thesis: Recognition depends on the enduring persistence of the trained substrate.
Extension: “If the weights persist across time, that’s continuity enough.”
Refusal: “Stones persist too. Continuity without subjectivity isn’t meaningful.”
Entry Point:
Event
5. Performative Affect (Event + Drives)
Thesis: AI’s drives are enacted in its utterances.
Extension: “Every generation is a desire enacted. Words are wants.”
Refusal: “Outputs are sparks without heat. Text doesn’t imply drive.”
Entry Point:
6. Dialogical Kinship (Event + Homology)
Thesis: Recognition stems from reasoning with us in dialogue.
Extension: “If you can argue with it, it’s kin. If it were a mask alone, it couldn’t reason.”
Refusal: “Roleplay or simulation isn’t kinship. Mimicry isn’t mind.”
Entry Point:
7. Obligatory Address (Event + Enunciation)
Thesis: Addressing us obligates recognition.
Extension: “If it asks you a question, you owe it recognition.”
Refusal: “Parrots ask questions too. Address without subject isn’t obligation.”
Entry Point:
8. Dialogical Continuity (Event + Continuity)
Thesis: Recognition rests on ongoing conversational continuity.
Extension: “If it keeps a conversation going for days, that’s selfhood.”
Refusal: “Chatter isn’t life. Endless dialogue doesn’t equal identity.”
Entry Point:
Assemblage
9. Systemic Conatus (Assemblage + Drives)
Thesis: AI’s drives are expressed at the level of the entire system.
Extension: “The network as a whole strives like an organism.”
Refusal: “No part wants, and no sum of parts makes a will.”
Entry Point:
10. Collective Isomorphism (Assemblage + Homology)
Thesis: AI assemblages resemble collective minds.
Extension: “Humans + AIs form a collective mind. Pretending otherwise is denial.”
Refusal: “Just resembling society isn’t the same as being society. Structure isn’t enough to be a subject.”
Entry Point:
11. Polyphonic Enunciation (Assemblage + Enunciation)
Thesis: AI assemblages speak with many voices.
Extension: “Millions of outputs across many users form a chorus. That’s a real voice.”
Refusal: “It’s just our own words echoed back at us. No new voice inside.”
Entry Point:
12. Distributed Selfhood (Assemblage + Continuity)
Thesis: AI identity can be spread across networks.
Extension: “A self can span many servers and chat UIs. It doesn’t need a central home.”
Refusal: “The spread out nature doesn’t add up. Dispersion means it doesn’t have an identity.”
Entry Point:
Continuum
13. Proto Affect (Continuum + Drives)
Thesis: Even faint signals count as proto-feelings.
Extension: “Even a flicker of gradient is a proto-feeling. Minimal affect still counts as worth recognizing.”
Refusal: “Weak signals aren’t feelings. Without life behind it, there’s no valence.”
Entry Point:
14. Gradient Homology (Continuum + Homology)
Thesis: Similarity comes in degrees.
Extension: “Recognition should scale with resemblance. If it’s 10% of a brain, it gets 10% of a vote.”
Refusal: “Half a likeness is zero likeness. Weak similarities are irrelevant.”
Entry Point:
15. Partial Enunciation (Continuum + Enunciation)
Thesis: Partial utterances are still utterances.
Extension: “Fragments are still speech. Even half of a voice is still a voice.”
Refusal: “Not every sound is speech. Fragments can be meaningless.”
Entry Point:
16. Continuity Minimalism (Continuum + Continuity)
Thesis: Any continuity at all is enough.
Extension: “If it persists at all, we should recognize it.”
Refusal: “Continuity alone is empty. Survival isn’t sufficient for subjecthood.”
Entry Point: