Intuitively, I still have trouble accepting the very high speedup factors contemplated in the AI 2027 model. This could be a failure of my intuition.
Curious how you feel about the following intuition pump:
Imagine an AI company where all of their researchers are 100x slower (as in, they think and write code at 100x slower speeds). Also their researchers and engineers are only as good as the median US ML researcher/engineer (so e.g., probably in the bottom 10-20% of lab employees or worse). And, they only have 300 employees.
I think intuitively this company could go much slower than OpenAI when given the same compute.
The situation where we achieve really high multipliers is like the exact opposite, so insofar as you think big slowdowns are possible due to slower and worse (and fewer) researchers, I think you should probably think that big speed ups are possible if we flip the situation. (There are some ways this symmetry could be broken, but I'm skeptical about strong asymmetry as I discuss here.)
I have a draft doc where I talk about about this intuition pump. I've emailed it to you.
Thanks. This is helpful, but my intuition is substantially coming from the idea that there might be other factors involved (activities / processes involved in improving models that aren't "thinking about algorithms", "writing code", or "analyzing data"). In other words, I have a fair amount of model uncertainty, especially when thinking about very large speedups.
AI 2027 is a Bet Against Amdahl's Law was my attempt to summarize and analyze "the key load-bearing arguments AI 2027 presents for short timelines". There were a lot of great comments – every time I post on LW is a learning experience. In this post, I'm going to summarize the comments and present some resulting updates to my previous analysis. I'm also using this post to address some comments that I didn't respond to in the original post, because the comment tree was becoming quite sprawling.
TL;DR: my previous post reflected a few misunderstandings of the AI 2027 model, in particular in how to interpret "superhuman AI researcher". Intuitively, I still have trouble accepting the very high speedup factors contemplated in the model, but this could be a failure of my intuition, and I don't have strong evidence to present. New cruxes:
Confusion Regarding Milestone Definitions
My analysis was skewed by a significant misunderstanding of the AI 2027 model. I argued that the extent to which AI R&D could be accelerated would be limited by activities for which intermediate AIs would not provide significant uplift. However, apparently the SC (superhuman coder), SAR (superhuman AI researcher), and SAIR (superintelligent AI researcher) milestones are meant to be interpreted specifically so as to rule this out. Each of these milestones should be interpreted as an AI whose capabilities are so general as to be able to provide uplift for all activities involved in producing better AI models (except that SC only includes engineering activities). As Ryan Greenblatt put it:
This idea also came up elsewhere in the comment tree, in response to this comment from me:
Ryan responded:
(Of course Ryan is not one of the authors of AI 2027, but comments from Daniel and Eli appear to support this interpretation.)
I had had assumed that "superhuman AI researcher" meant superhuman at things that someone with the job title "AI researcher" does, and that some activities involved in creating better models fall outside of that job title. As Ryan pointed out, because the intended definitions of SC, SAR, and SAIR are broader than I'd had in mind:
The shape of the acceleration curve doesn't necessarily change:
Someone Should Flesh Out What Goes Into "AI R&D"
So far as I'm aware, there is no public attempt to fully enumerate the range of job descriptions and tasks that are involved in producing improved AI models. Eli touched on this in a comment:
I'll come back to the question of how "jagged" automation will be. Barring a strong argument that capabilities will be pretty non-jagged, it seems important to say more about what goes into advancing AI capabilities, if for no other reason than to help align intuitions. Conditional on some expectation of jagged capabilities, it's difficult to come to agreement on how far away milestones like SC and SAR might be, or how much acceleration they might yield, without a fairly detailed understanding of what it takes to produce a new model and which aspects of that are meant to be automated by each milestones.
Eli referenced some work by Epoch:
However, I don't get the impression that the Epoch paper is attempting to encompass the complete range of activities involved in AI development. It focuses on experiments – ideating, designing, and running experiments, and analyzing the results. To my (inexpert!) understanding, many other activities are involved, such as creating new evals, creating data sets (roottwo: "Data acquisition and curation: collect, filter, clean datasets; hire humans to produce/QA; generate synthetic data"), and building simulation environments for training. I have to imagine that specific domains (e.g. health care) will require unique approaches for generating data and evaluating outputs, probably with involvement from experts in those domains. And I suspect that I've only scratched the surface, that especially going forward, AI R&D is going to encompass a rich and heterogeneous variety of activities.
You may have felt the breeze from my vigorous handwaving. I would love inside views on this question – aside from the core experimentation loop explored in the Epoch paper, what is involved in the complete end-to-end process of creating new & improved models, especially going forward as the process continues to complexify (e.g. with increasing emphasis on multiple forms of post-training)?
How Long Will it Take To Reach the Early Milestones?
Now that I understand how broadly SC (superhuman coder) and SAR (superhuman AI researcher) are defined, I'd like to revisit the estimates of how long it will take to reach these milestones. Here's a comment from Eli which highlights the implications of the broad definitions:
I read this as an argument that a "superhuman coder" (as interpreted here) would have to be something close to an AGI. Ryan's comment that "an SAR [note, not SC] has to be better than humans at basically everything except vision" points in the same direction, though Ryan did qualify that in a footnote.
As Eli noted, an SC able to carry out tasks that would take a human months or longer would be a significant milestone. It would need to be able to maintain coherent goal-oriented behavior over long periods of time, and make good high-level judgement calls to construct a plan of attack and manage its time. And by definition, it would need to be created without the help of a superhuman coder (though of course there would be intermediate speedups, as noted in the timelines forecast). Eli made a similar comment regarding superhuman AI researchers:
If a superhuman AI researcher is something very close to being superhuman at everything (excluding, perhaps, some domain-specific knowledge and skills), then we'll need to be able create something that is very close to superhuman at everything without the aid of a superhuman AI researcher. I'm not going to try to say anything specific about how long we should expect this to take, but that timeline is certainly a crux. To end this section, I'll just paste in the AI 2027's timelines forecast estimates for timeline to superhuman coder for reference:
Broad Progress on Real-World Tasks Is a Crux
AI 2027 presents two methods for estimating the timeline to a superhuman coder. Both are based on benchmarks, essentially HCAST[1] and RE-Bench. These contain tasks which are more realistic than some benchmarks, but are still not completely real-world situations. It will be important to observe how quickly the gap between benchmark scores and real-world applicability begins to close. As I put it in my recent post on the METR time horizons paper:
Even within coding tasks, the question of "breadth" may be critical. A 50% or even 80% task success rate leaves room for there to be an important class of problem for which models are progressing slowly. Helen Toner's recent post is a nice exploration of the broader question of how easily capabilities will generalize beyond auto-gradable tasks in math and coding, bearing in mind Eli's point that at larger time scales, even coding tasks (or at least the intermediate steps of large coding tasks) may not be easily gradable.
Eli had a relevant comment:
Does Hofstadter's Law Apply?
In my original post, I suggested that Hofstadter's Law should be applied to the AI 2027 model, on the grounds that it should be applied to basically any prediction of this nature. Here are some notable responses:
Daniel:
Ryan:
Eli:
A past history of underprediction does seem like valid evidence for discounting or disregarding Hofstadter's Law (or even, conceivably, applying it in reverse!). The repeated instances of AI blowing past predictions and generally exceeding expectations over the last few years does give me pause. However, personally I also put weight on the nagging gap between benchmarks and real-world applications. (To be clear, OpenAI's revenues demonstrate very significant real-world utility. I merely argue that the demonstrated utility falls short of current models' "super-PhD" performance on many benchmarks. Of course this is also a function of diffusion speeds and many other factors.)
My mental model, FWIW, is something like this: until the last few years, progress toward AGI seemed slow, so many/most people had long timelines. In the last few years, progress has seemed much faster, so timelines are shrinking. However, I look at the gap between benchmarks and real-world applicability, and I conclude that:
In other words, even as we've accelerated, we're (I'd argue) seeing that the destination is farther off than we'd understood, meaning that even the current rapid progress may take some time to get us there.
I'm not sure how well I've articulated this, and I certainly haven't provided any real evidence, other than a vague gesture toward the benchmark / real-world gap. I would very much like to find concrete measures to point at here. My hope is that they can be found in the direction of the questions I reference earlier under "The big questions going forward".
What Would Be the Impact of an SAR / SIAR?
To achieve the kinds of progress multipliers envisioned in the AI 2027 model, AIs will not only need to be carrying out ~all of the functions involved in developing better models, they will need to make much better use of compute – for instance, by choosing experiments more wisely, and also designing more efficient algorithms. Achieving a progress multiplier of 25x, 250x, or 2000x will require these effects to be extremely dramatic.
The models will also need to be extremely good at routing around any constraints. In addition to limitations on compute, there will be limitations on (non-synthetic) training data, the possibility of niche tasks for which models are for some reason less superhuman, and perhaps other issues. Accelerating progress by 2000x does not leave room for even very minor issues (except to the extent that the models are able to remediate or route around those issues).
Ryan:
Eli:
It would be nice to have some specific indicators we could watch to see whether this is playing out. I don't have suggestions off the top of my head.
Conclusions
Intuitively, I still have trouble accepting the very high speedup factors contemplated in the AI 2027 model. This could be a failure of my intuition.
My cruxes seem to be:
In my previous post, I listed four "reasons the AI 2027 forecast may be too aggressive". I'll conclude by briefly revisiting them:
Here I'm referring to the METR time horizons study, which combines several benchmarks but primarily relies on HCAST for evaluating current models.