Some people have expressed that “GPT-2 doesn’t understand anything about language or reality. It’s just huge statistics.” In at least two senses, this is true.
First, GPT-2 has no sensory organs. So when it talks about how things look or sound or feel and gets it right, it is just because it read something similar on the web somewhere. The best understanding it could have is the kind of understanding one gets from reading, not from direct experiences. Nor does it have the kind of understanding that a person does when reading, where the words bring to mind memories of past direct experiences.
Second, GPT-2 has no qualia. This is related to... (read 1357 more words →)
I'm not sure this constraint (single forward pass as opposed to reasoning-enabled) is very meaningful. If it is still doing the reasoning as it answers the question, does it make a big difference whether that is enclosed by <thinking></thinking> tags? I would be interested to see a version of this where the answer has to be the first token output. That would show how much thinking it can do in one step.