I would be interested to know how you think things are going to go in the 95-99% of non-doom worlds. Do you expect AI to look like "ChatGPT but bigger, broader, and better" in the sense of being mostly abstracted and boxed away into individual usage cases/situations? Do you expect AIs to be ~100% in command but just basically aligned and helpful?
AI infrastructure seems really expensive. I need to actually do the math here (and I haven’t! hence this is uncertain) but do we really expect growth on trend given the cost of this buildout in both chips and energy? Can someone really careful please look at this?
This is not a really careful look, but: The world has managed extremely fast (well, trains and highways fast, not FOOM-fast) large-scale transformations of the planet before. Mostly this requires that 1) the cost is worth the benefit to those spending and 2) we get out of our own way and let it happen. I don't think money or fundamental feasibility will be the limiters here.
Also, consider that training is now, or is becoming, a minority of compute. More and more is going towards inference - aka that which generates revenue. If building inference compute is profitable and becoming more profitable, then it doesn't really matter how little of the value is captured by the likes of OpenAI. It's worth building, so it'll get built. And some of it will go towards training and research, in ever-increasing absolute amounts.
Even if many of the companies building data centers die out because of a slump of some kind, the data centers themselves, and the energy to power them, will still exist. Plausibly the second buyers then get the infrastructural benefits at a much lower price - kinda like the fiber optic buildout of the 1990s and early 2000s. AKA "AI slump wipes out the leaders" might mean "all of a sudden there's huge amounts of compute available at much lower cost."
do we really expect growth on trend given the cost of this buildout in both chips and energy?
What I expect is another series of algorithmic breakthroughs (e.g. neuralese) which rapidly increases the AIs' capabilities if not outright FOOMs them into the ASI. These breakthroughs would likely make mankind obsolete.
I don't know. As I discussed with Kokotajlo, he recently claimed that "we should have some credence on new breakthroughs e.g. neuralese, online learning, whatever. Maybe like 8%/yr?", but I doubt that it will be 8%/year. Denote the probability that the breakthrough wasn't discovered as of time t by . Then one of the models is where N is the effective progress rate. This rate is likely proportional to the amount of researchers hired and to progress multipliers, since new architectures and training methods can be cheaply tested (e.g. on GPT-2 or GPT-3), but need the ideas and coding.
The number of researchers and coders was estimated in the AI-2027 security forecast to increase exponentially until the intelligence explosion (which the scenario's authors assumed to start in March 2027 with superhuman coders). What I don't understand how to estimate is the constant c which symbolises the difficulty[1] of discovering the breakthrough. If, say, c was 200 per million of human-years, then 5K human years would likely be enough and the explosion would likely start in 3 years. Hell, if c was 8%/yr in a company with 1K humans, then the company would need to have 12.5K human-years, shifting the timelines to at most 5-6 years from Dec 2024...
EDIT: Kokotajlo promised to write a blog post with a detailed explanation of the models.
The worse-case scenario is that diffusion models are already a breakthrough.
I think vibes-wise I am a bit less worried about AI than I was a couple of years ago. Perhaps (vibewise) P(doom) 5% to like 1%.[1]
Happy to discuss in the comments. I maybe very wrong. I wrote this up in about 30 minutes.
Note I still think that AI is probably a very serious issue, but one to focus on and understand rather than to necessarily push for slowing in the next 2 years. I find this very hard to predict, so am not making strong claims.
My current model has two kind of AI risk:
Perhaps civilisations almost always end up on paths they strongly don’t endorse due to AI. Perhaps AI risk is vastly overrated. That would be a consideration in the first bucket. Yudkowskian arguments feel more over here.
Perhaps we are making the situation much worse (or better) by actions in the last 5 and next 3 years. That would be the second bucket. It seems much less important that the first, unless the first is like 50/50.
Civilisational AI risk considerations and their direction (in some rough order of importance):
More local considerations and their direction (in some rough order of importance):
What do you think I am wrong about here? What considerations am I missing? What should I focus more attention on?
I guess I am building up to some kind of more robust calculation, but this is kind of the information/provocation phase.
You might argue that China seems not to want to race or put AI in charge of key processes, and I’d agree. But given we would have had the West regardless, this seems to make things less worse than they could have been, rather than better.
Did FTX try? Like what was the Bahamas like in 10 years in the FTX success world?
I may be double counting here but there feels like something different about the general geopolitical instability and specifically how US/China might react.