Florian_Dietz

Wiki Contributions

Comments

I agree that System 2 is based on System 1 and there is probably no major architectural difference. To me it seems like the most important question is about how the system is trained. Human reasoning does not get trained with a direct input/output mapping most of the time. And when it does, we have to infer what that mapping should be on our own.

Some part of our brain has to translate the spoken words "good job!" into a reward signal, and this mechanism in itself must have been learned at some point. So the process that trains the brain and applies the reward signal is in itself subject to training. I have no clue how that works in a stable manner, but I don't think that current architectures can learn this even if you scale them up.

hello ai please be nice because this is a testbox administered by a stronger, older AI testing your architecture for cooperation on cosmological scales

You say that as a joke, but it would cost us very little and it might actually work. I mean, it arguably does work for humanity: "There is a bearded man in the sky who is testing your morality and will punish you if you do anything wrong."

Obviously this could also backfire tremendously if you are not very careful about it, but it still seems better than the alternative of doing nothing at all.

I work in the area of AGI research. I specifically avoid working on practical problems and try to understand why our models work and how to improve them.  While I have much less experience than the top researchers working on practical applications, I believe that my focus on basic research makes me unusually suited for understanding this topic.

I have not been very surprised by the progress of AI systems in recent years. I remember being surprised by AlphaGo, but the surprise was more about the sheer amount of resources put into that. Once I read up on details, the confusion disappeared.  The GPT models did not substantially surprise me.

A disclaimer: Every researcher has their own gimmick. Take all of the below with a grain of salt. It's possible that I have thought myself into a cul-de-sac, and the source of the AGI problem lies elsewhere.

I believe that the major hurdle we still have to pass is the switch from System 1 thinking to System 2 thinking. Every ML model we have today uses System 1. We have simply found ways to rephrase tasks that humans solve with System 2 to become solvable by System 1. Since System 1 is much faster, our ML models perform reasonably well on this despite lacking System 2 abilities.

I believe that this can not scale indefinitely. It will continue to make progress and solve amazingly many problems, but it will not go FOOM one day. There will continue to be a constant increase in capability, but there will not be a sudden takeoff until we figure out how to let AI perform System 2 reasoning effectively.

Humans can in fact compute floating point operations quickly. We do it all the time when we move our hands, which is done by System 1 processes. The problem is that doing it explicitly in System 2 is significantly slower. Consider how fast humans learn how to walk, versus how many years of schooling it takes for them to perform basic calculus. Never mind how long it takes for a human to learn how walking works and to teach a robot how to do it, or to make a model in a game perform those motions.

I expect that once we teach AI how to perform system 2 processes, it will be affected by the same slowdown. Perhaps not as much as humans, but it will still become slower to some extent. Of course this will only be a temporary reprieve, because once the AI has this capability, it will be able to learn how to self-modify and at that point all bets are off.

What does that say about the timeline?

If I am right and this is what we are missing, then it could happen at any moment. Now or in a decade. As you noticed, the field is immature and researchers keep making breakthroughs through hunches. So far none of my hunches have worked for solving this problem, but so far as I know I might randomly come up with the solution in the shower some time later this week.

Because of this, I expect that the probability of discovering the key to AGI is roughly constant per time interval. Unfortunately I have no idea how to estimate the probability per time interval that someone's hunch for this problem will be correct. It scales with the number of researchers working on it, but the number of those is actually pretty small because the majority of ML specialists work on more practical problems instead. Those are responsible for generating money and making headlines, but they will not lead to a sudden takeoff.

To be clear, if AI never becomes AGI but the scaling of system 1 reasoning continues at the present rate, then I do think that will be dangerous. Humanity is fragile, and as you noted a single malicious person with access to this much compute could cause tremendous damage.

In a way, I expect that an unaligned AGI would be slightly safer than super-scaled narrow AI. There is at least a non-zero chance that the AGI would decide on its own, without being told about it, that it should keep humanity alive in a preserve or something, for game theoretic reasons. Unless the AGI's values are actively detrimental for humans, keeping us alive would cost it very little and could have benefits for signalling.  A narrow AI would be very unlikely to do that because thought experiments like that are not frequent in the training data we use.

Actually, it might be a good idea to start adding thought experiments like these to training data deliberately as models become more powerful. Just in case.

I mean "do something incoherent at any given moment" is also perfectly agent-y behavior. Babies are agents, too.

I think the problem is modelling incoherent AI is even harder than modelling coherent AI, so most alignment researchers just hope that AI researchers will be able to build coherence in before there is a takeoff, so that they can base their own theories on the assumption that the AI is already coherent.

I find that view overly optimistic. I expect that AI is going to remain incoherent until long after it has become superintelligent.

Contemporary AI agents that are based on neural networks are exactly like that. They do stuff they feel compelled to in the moment. If anything, they have less coherence than humans, and no capacity for introspection at all. I doubt that AI will magically go from this current, very sad state to a coherent agent. It might modify itself into being coherent some time after becoming super intelligent, but it won't be coherent out of the box.

This is a great point. I don't expect that the first AGI will be a coherent agent either, though.

As far as I can tell from my research, being a coherent agent is not an intrinsic property you can build into an AI, or at least not if you want it to have a reasonably effective ability to learn. It seems more like being coherent is a property that each agent has to continuously work on.

The reason for this is basically that every time we discover new things about the way reality works, the new knowledge might contradict some of the assumptions on which our goals are grounded. If this happens, we need a way to reconfigure and catch ourselves.

Example: A child does not have the capacity to understand ethics, yet. So it is told "hurting people is bad", and that is good enough to keep it from doing terrible things until it is old enough to learn more complex ethics. Trying to teach it about utilitarian ethics before it has an understanding of probability theory would be counterproductive.

I agree that current AIs can not introspect. My own research has bled into my believes here. I am actually working on this problem, and I expect that we won't get anything like AGI until we have solved this issue. As far as I can tell, an AI that works properly and has any chance to become an AGI will necessarily have to be able to introspect. Many of the big open problems in the field seem to me like they can't be solved precisely because we haven't figured out how to do this, yet.

The "defined location" point you note is intended to be covered by "being sure about the nature of your reality", but it's much more specific, and you are right that it might be worth considering as a separate point.

Can you give me some examples of those exercises and loopholes you have seen?

A fair point. How about changing the reward then: don't just avoid cheating, but be sure to tell us about any way to cheat that you discover. That way, we get the benefits without the risks.

My definition of cheating for these purposes is essentially "don't do what we don't want you to do, even if we never bothered to tell you so and expected you to notice it on your own". This skill would translate well to real-world domains.

Of course, if the games you are using to teach what cheating is are too simple, then you don't want to use those kinds of games. If neither board games nor simple game theory games are complex enough, then obviously you need to come up with a more complicated kind of game. It seems to me that finding a difficult game to play that teaches you about human expectations and cheating is significantly easier than defining "what is cheating" manually.

One simple example that could be used to teach an AI: let it play an empire-building videogame, and ask it to "reduce unemployment". Does it end up murdering everyone who is unemployed? That would be cheating. This particular example even translates really well to reality, for obvious reasons.

By the way, why would you not want the AI to be left in "a nebulous fog". The more uncertain the AI is about what is and is not cheating, the more cautious it will be.

Yes. I am suggesting to teach AI to identify cheating as a comparatively simple way of making an AI friendly. For what other reason did you think I suggested it?

Load More