[ Question ]

Are we certain that gpt-2 and similar algorithms are not self-aware?

by Ozyrus 1 min read11th Jul 20199 comments


Have someone even started this conversation? This is f*cked up. I'm really, really freaked out lately with some of those. https://old.reddit.com/r/SubSimulatorGPT2/comments/cbauf3/i_am_an_ai/

And I read up a lot about cognition and AI, and I'm pretty certain that they are not. But shouldn't we give it a lot more consideration just because there's a small chance they are? Because the ramifications are that dire.

But then again, I'm not sure that knowledge will help us in any way.

New Answer
Ask Related Question
New Comment
Write here. Select text for formatting options.
We support LaTeX: Cmd-4 for inline, Cmd-M for block-level (Ctrl on Windows).
You can switch between rich text and markdown in your user settings.

1 Answers

Yes, we are.

Wikipedia defines self-awareness as "the capacity for introspection and the ability to recognize oneself as an individual separate from the environment and other individuals". At a minimum, this would require GPT-2 to have a model of the world which included a representation of itself. To be similar to our intuitive understanding of self-awareness, that representation would also need to guide its decision-making and thought in some significant way.

Here is an intuitive explanation of the Transformer architecture that GPT-2 is based on. You can see from the explanation that it's only modeling language; there's no self-representation involved.

Technically, I guess you could say that, if a Transformer architecture was trained on texts which talked about Transformer architectures, it would get a model which did include a representation of itself. But that would be just another data token, which the system gave no special significance to, and which wouldn't guide its behavior any more than any other piece of data.