It feels to me like Sutton is too deep inside the experiential learning theory. When he says there is no evidence for imitation, this only makes sense if you imagine it strictly according to the RL theory he has in mind. He isn't applying the theory to anything; he is inside the theory and interpreting everything according his understanding of it.
It did feel like there was a lot of talking past one another when Dwarkesh was clearly talking about the superintelligent behaviors everyone is interested in (doing science, math, and engineering) as his model for intelligence, and Sutton is blowing all of this off only to articulate quite late in the game that his perspective is that human infants are his model for intelligence. If this was cleared up early, it would probably have been more productive.
I have always found the concept of a p-zombie kind of silly, but now I feel like we might really have to investigate the question of an approximate i-zombie: if we have a computer than can output anything an intelligent human can, but we stipulate that the computer is not intelligent....and so on and so forth.
On the flip side, it feels kind of like a waste of time. Who would be persuaded by such a thing?
Obviously the training data of LLMs contains more than human dialogue, so the claim that the pretrained LLMs are "strictly imitating humans" is clearly false. I don't know why this was never brought up.
It's neither obvious nor clear to me. Who wrote the rest of their training data, besides us oh-so-fallible humans? What percentage of the data does this non-human authorship constitute?
Modern multimodal LLMs are trained not just on text data, but also on images and video. In the finetuning stage for reasoning models, they are trained not to predict human reasoning, but to do reasoning themselves to accomplish a binary goal. Not sure about exact percentages but in terms of capabilities, they can recognize images very well and generate very realistic images and videos, to the point where many people can't tell them apart.
That gives them more different abilities; I don't think it constitutes a fundamental change to their way of thinking or that it makes them more intelligent.
(It doesn't improve their performance on text based problems significantly.)
Because it is just doing the ~same type of "learning" on a different type of data.
This doesn't make them able to discuss say abiogenesis or philosophy with actual critical human-like thought. In these fields they are strictly imitating humans.
As in, imagine you replaced all the learning data regarding abiogenesis with plausible-sounding but subtly wrong theories. The LLM would simply slavishly repeat these wrong theories, wouldn't it?
Sutton seems to confuse intelligence with life. These are distinctly different concepts. Compare LLMs and bacteria: LLMs are intelligent but not alive, bacteria are alive but not intelligent. Bacteria have goals, such as consuming food and avoiding hazards, and bacteria take directed action to accomplish their goals.
John McCarthy’sdefinition that intelligence is the computational part of the ability to achieve goals
.....needs to be taken broadly -- the goals don't have to be the intelligent agents own. An intelligent servant is quite conceivable
Probably the most sensible response to the interview I've seen so far.
Also I'll probably start referencing this when people argue whether thinking in active inference frame has any advantage over thinking in the RL frame. Clearly it does: it's way easier to see what happens if you drop the "reward" term Sutton is imagining as necessary and keep just the prediction error minimization terms. You still get intelligent systems, they still learn powerful abstractions (because they need to compress data), they still learn a generative world model. (It's probably good the active inference frame is antimemetic in the orthodox RL crowd)
- Almost all feedback is noisy because almost all outcomes are probabilistic.
Yes but signal / noise ratios matter a lot.
Language is somewhat optimized to pick up signal and skip noise. For example "red" makes it easier to pick ripe fruit, "grue" doesn't really exist because its useless, "expired" is a real concept because it's useful.
It also has some noise added. For example putting (murderers and jay wakers) in a category "criminal" to politically oppose something.
Also not being exposed to the kind of noise that's present IRL might be an issue when you start to deal with IRL (sometimes people say something like "just do the max EV action" is a good enough plan)
I'm pretty sure this is some obstacle for LLMs, I'm pretty sure its something that can be overcome, I'm very unsure how much this matters.
Continues Learning is indeed a big deal.
Continues learning does not mean you need a continuously learning LLM - this can be a property of a system.
But:
This seems like a good opportunity to do some of my classic detailed podcast coverage.
The conventions are:
Full transcript of the episode is here if you want to verify exactly what was said.
Well, that was the plan. This turned largely into me quoting Sutton and then expressing my mind boggling. A lot of what was interesting about this talk was in the back and forth or the ways Sutton lays things out in ways that I found impossible to excerpt, so one could consider following along with the transcript or while listening.
Sutton Says LLMs Are Not Intelligent And Don’t Do Anything
Sutton has a reasonable hypothesis that a different architecture, that uses a form of continual learning and that does so via real world interaction, would be an interesting and potentially better approach to AI. That might be true.
But his uses of words do not seem to match their definitions or common usage, his characterizations of LLMs seem deeply confused, and he’s drawing a bunch of distinctinctions and treating them as meaningful in ways that I don’t understand. This results in absurd claims like ‘LLMs are not intelligent and do not have goals’ and that feedback from digital systems doesn’t count and so on.
It seems like a form of essentialism, the idea that ‘oh LLMs can never [X] because they don’t [Y]’ where when you then point (as people frequently do) to the LLM doing [X] and often also doing [Y] and they say ‘la la la can’t hear you.’
Humans Do Imitation Learning
The Experimental Paradigm
Current Architectures Generalize Poorly Out Of Distribution
So do humans.
Surprises In The AI Field
Will The Bitter Lesson Apply After AGI?
Succession To AI
It’s Sutton, so eventually we were going to have to deal with him being a successionist.
Fundamentally, I didn’t pull direct quotes on this but Sutton repeatedly emphasizes that AI-dominated futures can be good or bad, that he wants us to steer towards good futures rather than bad futures, and that we should think carefully about which futures we are steering towards and choose deliberately.
I can certainly get behind that. The difference is that I don’t think we need to accept this transition to AI dominance as our only option, including that I don’t think we should accept that humans will always be unable to coordinate.
Mostly what I found interesting were the claims around the limitations and nature of LLMs, in ways that don’t make sense to me. This did help solidify a bunch of my thinking about how all of this works, so it felt like a good use of time for that alone.