Epistemic Status: Exploratory
I wrote this as part of an application for the Chicago Symposium on Transformative AI, where I try & sketch out what takeoff might look like. I’m making a lot of claims, across a range of domains, so I’d expect there to be many places I’m wrong. But on the whole, I hope this is more well-thought-out than not.[1]
Many thanks to Nikola Jurković & Tao Burga for thoughtful comments on this writeup. Any errors are, of course, my own.
Prompt: Please outline an AGI takeoff scenario in detail, noting your key uncertainties
If it’s closer to the latter, I think we’re doomed. Why? Tl;dr: too much optimization pressure is bad; and if we crank it up high enough, we goodhart ourselves out of existence.[2]
That’s not a very nice scenario to analyze. So I’m focusing on worlds in which we achieve the former & not the latter.[3]
Ie. as opposed to multiple actors converging upon AGI simultaneously. Why? I think it’s far likelier that we ‘stumble onto’ AGI – given that progress has often been empirically (and not theory) driven.[4]
I’ve updated away from 3-year timelines with the release of GPT 4.5. Primarily, it seems like we’re getting diminishing returns from pre-training compute.[5]
For one, DeepSeek made a lot of efficiency gains and open-sourced all their research. Labs will catch up with these gains shortly. While this might not raise the floor of capabilities, it will make ‘access to intelligence’ much cheaper, and likely speed up internal progress at labs.[6]
Also, RL is incredibly powerful, and we’re only starting to harness it with today’s reasoning models. It doesn’t seem like we’ve got a principled understanding of how to do RL well, and it seems like there’s a fairly high ceiling here.[7]
It seems like internal compartmentalization will be hard.[8] Once enough employees get wind of it, it’s likely that one of them will blow the whistle.
There might also be other signs of increased intensity – later-than-usual hours, public absence of top researchers, etc.[9]
First time posting here – any feedback is appreciated!
Or at least, that’s the default outcome with our progress as it stands today. I think that’s what folks are worried about with the sharp left turn.
Or at least, there’s some moderate amount of time before we go from labor-automators to paperclip-maximizers.
This isn’t a claim about when this would happen / how much compute would be required / etc. But it will likely be unexpected, whenever it may be.
I think this post raises some good points.
I think Gwern had a good take on this.
I’m uncertain about this claim. I don’t have in-depth knowledge, but this is my impression.
Labs still have to run evals / implement safety measures / etc – ie. perform tasks that need contact time with the system.
I don’t have many plausible examples of these, but they seem pretty likely on my inner sim. Not very confident about this point, and would love to hear other thoughts.