I've got a new working paper, title above, abstract, contents, introduction, and conclusion below.

* * * * *

Abstract: I make three arguments. There is a philosophical argument that the behavior of ChatGPT is so sophisticated that the ordinary concept of thinking is no longer useful in distinguishing between human behavior and the behavior of advanced AI. We don’t have deep and explicit understanding about what either humans or advanced AI systems are doing. The other argument is about ChatGPT’s behavior. As a result of examining its output in a systematic way, short stories in particular, I have concluded that its operation is organized on at least two levels: 1) the parameters and layers of the LLN, and 2) higher level grammars, if you will, that are implemented in those parameters and layers. This is analogous to the way that high-level programming languages are implemented in assembly code. 3) Consequently, it turns out that aspects of symbolic computation are latent in LLMs. An appendix gives examples of how a story grammar is organized into frames, slots, and fillers.

Contents

0. What is in here and why ........................................................................................ 2
1. The Chinese Room is empty  ................................................................................ 3
2. ChatGPT, justice (an abstract idea), and interpretation (analogy)  .............. 4
3. Tell me a story ........................................................................................................  6
4. Here be dragons: The idea of thought is broken  ........................................... 10
5. What underlying structure drives ChatGPT’s behavior? ............................. 11
6. Where are we?  ....................................................................................................  14
7. Appendix 1: ChatGPT derives new stories from old  ................................... 16
8. Appendix 2: Texts considered as 1-dimensional visual objects  ................. 21
9. Appendix 3: My work about ChatGPT  ............................................................21

0. What is in here and why

After having worked with ChatGPT intensely for a month and a half I am all but convinced that the underlying language model is organized on at least two levels: 1) There is a “bottom” level consisting of 175 billion weighted parameters organized into layers. This structure is generally considered opaque. 2) There is a higher level that seems all but invisible; the bottom level implements this higher level and it is this higher level that organizes discourse with users. We can infer this higher level by analyzing the structure of texts. I have been analyzing and describing (some of) the manifestations of this higher level as a necessary precursor to characterizing it as a “grammar” or computational structure.

That is one argument I make. I interleave it with another one, that ChatGPT is so successful in enacting a rich and varied simulacrum of human behavior that philosophical arguments asserting that computers cannot, in principle, think, such arguments are rapidly losing their cogency. They may remain valid, but they tell us little about what humans can do that computers can’t.

* * * * *

1. Why the Chinese Room is empty – Searle’s argument, and others like it, got much of its rhetorical force from the fact that AI systems at the time were quite limited in their capacity. That is no longer the case. Those arguments are rapidly losing their potency.

2. ChatGPT, justice (an abstract idea), and interpretation (analogy) – Analogical reasoning, in the form of interpreting texts (movies) and reasoning about abstract concepts (justice).

3. Tell me a story – I use a procedure derived from Lévi-Strauss’s work on myth to probe GPT’s capacity for story-telling.

4. Here be dragons: The idea of thought is broken – Advanced AI is unlike anything we’ve seen before. These systems are not human, are not sentient, yet their behavior appears human. We need concepts (with new terms) for understanding them.

5. What underlying structure drives ChatGPT’s behavior? – There is the neural net, yes. But the neural net serves as a medium for implementing higher level structures, as assembly languages are a medium for implementing high-level programming languages.

6. Where are we? – A reprise of sections 1 through 5 ending with the suggestion that symbolic structures are latent in ChatGPT.

7. Appendix 1: ChatGPT derives new stories from old – Tables associated with the story mechanisms described in section 3.

8. Appendix 2: Texts considered as 1-dimensional visual objects – A thought experiment in which texts are transformed into strings of colored beads.

9. Appendix 3: My work about ChatGPT – Links to my work, blog posts, working papers, on ChatGPT, GPT, and deep learning.

* * * * *

6. Where are we?

I read Searle’s Chinese Room argument when it was first published in Brain and Behavioral Science. I found it most peculiar and unsatisfactory for it didn’t engage with any of the techniques that were then being used in artificial intelligence or computational linguistics. What use is an argument about computers and mind if it doesn’t tell me anything I can use in designing a computational mind? At the time I was still deeply engaged with the work I had done with cognitive networks in graduate school.

In the past decade or so I have come to appreciate the force of Searle’s argument, but I have also come to realize that much of that force came, not from the argument itself, but from the intellectual context in which it arose. The AI systems of the time weren’t that impressive. Their behavior did not present phenomenological evidence that they were thinking. That is no longer the case. Advanced AI systems exhibit such powerful phenomenological evidence that many are willing to declare, “Yes, they’re thinking.”

Maybe they are, maybe they aren’t. On the whole, I think not. But purely philosophical arguments are no longer a satisfactory source of insight into the nature of thought, whether in humans or machines. The concept of thought is broken. It needs to be reconstructed on technical and empirical grounds, in terms applicable both to computers and to brains. We need new concepts and new terminology.

One technical point at issue is the question of how large language models work. I have been arguing that the model driving ChatGPT needs to be understood on at least two levels. There is the level of parameters and layers in the network, which is considered to be opaque. That is where almost all current efforts are focused.

I have been arguing that that work, no matter how successful, isn’t enough. You can’t get there from here. Just as a computer language is a platform on which applications can be constructed, so an artificial neural net is a platform on which artificial “minds,” if you will, can be constructed. By analyzing ChatGPT’s linguistic behavior using terms we would use in analyzing human linguistic behavior, we see recurring patterns that must be accounted for. I have investigated those recurring patterns in a variety of ways, perhaps most intensively for stories, which I have concentrated on in this paper. I have argued that the the regularities in the stories ChatGPT tells are best accounted for by some kind of story grammar. That story grammar is, in turn, implemented in a large language model.

It seems natural to me to analyze ChatGPT’s linguistic output as I would texts produced by humans. I am certaintly not the only one to do so. I have already cited that work that Taylor Webb et al. have done investigating GPT-3’s capacity for analogical reasoning. They used test instruments originally designed for assessing human capabilities. Marcel Binz and Eric Schulz have recently called for using the tools of cognitive psychology in assessing GPT-3’s performance.

It is true that real nervous systems and artificial neural nets are different in many ways, but we know that both can handle a wide variety of perceptual and cognitive tasks. At some ‘level’ the model must necessarily be dominated by the structure of the problem domain – in this case, strings of word forms – rather than the inherent properties of the modeling matrix. What you want from such a matrix is the flexibility to adapt to the demands of a wide variety of problem domains. And that’s what we see, certainly in the case of real nervous systems, but in the case of ANNs as well.

Thus one would expect that the kinds of conceptual instruments developed for understanding human story-telling would also be useful in understanding stories told by ChatGPT. I note that such grammars have been well investigated in the era of symbolic computing. That suggests, in turn, that the question of whether or not symbolic computing is necessary to achieving the full potential of artificial intelligence, that question is malformed. Large language models already seem to be using symbolic mechanisms even though they were not designed to do so. Those symbolic mechanisims have simply emerged, albeit implicitly. They are latent in the system. What will it take to make them explicit?

We do not want to hand-code symbolic representations, like researchers did in the GOFAI era. The universe of discourse is too large and complex for that. Rather we need some way to take a language language model and bootstrap explicit symbolic representations into – or is it over? – it. It is one thing to “bolt” a symbolic system onto neural net. How do we get the symbolic system to emerge from the neural net, as it does in humans?

During the first year and a half of life, children acquire a rich stock of ‘knowledge’ about the physical world and about interacting with others. They bring that knowledge with them as their interaction with others broadens to include language. They learn language by conversing. In this way word forms become indexes into their mental model of the world.

Can we figure out a similar approach for artificial systems? They can converse with us, and with each other. How do we make it work?

6

New Comment

New to LessWrong?