https://www.elilifland.com/. You can give me anonymous feedback here. I often change my mind and don't necessarily endorse past writings.
Thanks for writing this up, glad to see the engagement! I've only skimmed and have not run this by any other AI 2027 authors, but a few thoughts on particular sections:
My predictions for AI by the end of 2027
I agree with most but not all of these in the median case, AI 2027 was roughly my 80th percentile aggressiveness prediction at the time.
Edited to add, I feel like I should list the ones that I have <50% on explicitly:
AI still can't tell novel funny jokes, write clever prose, generate great business ideas, invent new in-demand products, or generate important scientific breakthroughs, except by accident.
I disagree re: novel funny jokes, seems plausible that this bar has already been passed. I agree with the rest except maybe clever prose, depending on the operationalization.
LLMs are broadly acknowledged to be plateauing, and there is a broader discussion about what kind of AI will have to replace it.
Disagree but not super confident.
Most breakthroughs in AI are not a result of directly increasing the general intelligence/"IQ" of the model, e.g. advances in memory, reasoning or agency. AI can stay on task much longer than before without supervision, especially for well-specified, simple tasks. Especially since AI coding platforms will have gotten better at tool use and allowing AI to manually test the thing they're working on. By the end of 2027, AI can beat a wide variety of video games is hasn't played before.
I disagree with the first clause, but I'm not sure what you mean because advances in reasoning and agency seem to me like examples of increases in general intelligence. Especially staying on task for longer without supervision. Are you saying that these reasoning and agency advances will mostly come from scaffolding rather than the underlying model getting smarter? That I disagree with.
There is more public discussion on e.g. Hacker News about AI code rot and the downsides of using AI. People have been burned by relying too much on AI. But I think non-coders running businesses will still by hyped about AI in 2027.
Disagree on the first two sentences.
AI still can't drive a damned car well enough that if I bought a car I wouldn't have to.
I don't follow self-driving stuff much, but this might depend on location? Seems like good self-driving cars are getting rolled out in limited areas at the moment.
As you touch on later in your post, it's plausible that we made a mistake by focusing on 2027 in particular:
But I do worry about what happens in 2028, when everyone realizes none of the doomsday stuff predicted in 2025 actually came true, or even came close. Then the AI alignment project as a whole may risk being taken as seriously as the 2012 apocalypse theory was in 2013. The last thing you want is to be seen as crackpots.
I think this is a very reasonable concern and we probably should have done better in our initial release making our uncertainty about timelines clear (and/or taking the time to rewrite and push back to a later time frame, e.g. once Daniel's median changed to 2028). We are hoping to do better on this in future releases, including via just having scenarios be further out, and perhaps better communicating our timelines distributions.
Also:
Listening to several of the authors discuss the AI 2027 predictions after they were published leads me to believe they don't intuitively believe their own estimates.
What do you mean by this? My guess is that it's related to the communciation issues on timelines?
The Takeoff Forecast is Based on Guesswork
Agree.
The Presentation was Misleading
Nothing wrong with guesswork, of course, if it's all you've got! But I would have felt a lot better if the front page of the document had said "AI 2027: Our best guess about what AI progress might look like, formulated by using math to combine our arbitrary intuitions about what might happen."
But instead it claims to be based on "trend extrapolations, wargames, expert feedback, experience at OpenAI, and previous forecasting successes", and links to 193 pages of data/theory/evidence.
They never outright stated it wasn't based on vibes, of course, and if you dig into the document, that's what you find out.
I very much understand this take and understand where you're coming from because it's a complaint I've had regarding some previous timelines/takeoff forecasts.
Probably some of our disagreement is very tied-in to the object-level disagreements about the usefulness of doing this sort of forecasting; I personally think that although the timelines and takeoff forecasts clearly involved a ton of guesswork, they are still some of the best forecasts out there, and we need to base our timelines and takeoff forecasts on something in the absence of good data.
But still, since we both agree that the forecasts rely on lots of guesswork, even if we disagree on their usefulness, we might be able to have some common ground when discussing whether the presentation was misleading in this respect. I'll share a few thoughts from my perspective below:
Linking to some other comments I've written that are relevant to this: here, here
In terms of general intelligence including long-horizon agency, reliability, etc., do we think AIs are yet, for example, as autonomously good as the worst professionals? My instinct is no for many of them, even though the AIs might be better at the majority of sub-tasks and are very helpful as collaborators rather than fully replacing someone. But I'm uncertain, it might depend on the operalization and profession, for some professions the answer seems clearly yes.[1][2] It also seems harder to reason about than the literally least capable professional something like the 10th percentile.
If the answer is no and we're looking at the ability to fully autonomously replace humans, this would mean the village idiot -> Einstein claim might technically not be falsified. The spirit of the claim might be though, e.g. in terms of the claimed implications.
There's also a question of whether we should include phyiscal abilities, if so then the answer would clearly be no for those professions or tasks.
One profession for which it seems likely that the AIs are better than the least capable humans is therapy. Also teaching/tutoring. In general this seems true for professions that can be done via remote work, don't involve heavy required computer use or long horizon agency.
I'd be excited for people (with aid of LLMs) to go back and grade how various past predictions from MIRI folks are doing, plus ideally others who disagreed. I just read back through part of https://www.lesswrong.com/posts/vwLxd6hhFvPbvKmBH/yudkowsky-and-christiano-discuss-takeoff-speeds and my quick take is that Paul looks mildly better than Eliezer due to predicting larger impacts/revenue/investment pre-AGI (which we appear to be on track for and to some extent already seeing) and predicitng a more smooth increase in coding abilities, but hard to say in part because Eliezer mostly didn't want to make confident predictions, also I think Paul was wrong about Nvidia but that felt like an aside.
edit: oh also there's the IMO bet, I didn't get to that part on my partial re-read, that one goes to Eliezer.
Looking through IEM and the Yudkowsky-Hanson debate also seems like potentially useful sources, as well as things that I'm probably forgetting or unaware of.
Eli was saying something like "90%-horizons of 100 years sound about right for Superhuman Coder level performance"
To be clear, this is on (a theoretical extrapolated version of) METR HCAST, not the real world distribution of software engineering projects.
Also to remind others of the definition of superhuman coder, it's a pretty high bar:
Superhuman coder (SC): An AI system for which the company could run with 5% of their compute budget 30x as many agents as they have human research engineers, each of which is on average accomplishing coding tasks involved in AI research (e.g. experiment implementation but not ideation/prioritization) at 30x the speed (i.e. the tasks take them 30x less time, not necessarily that they write or “think” at 30x the speed of humans) of the company’s best engineer. This includes being able to accomplish tasks that are in any human researchers’ area of expertise.
Yeah this is the primary argument pushing me toward thinking there shouldn't be a finite-time singularity, as I mentioned I'm not confident. It does feel pretty crazy that a limits-of-intelligence ASI would have a (very large horizon) time horizon at which it has 0.00001% reliability though, which I think is unavoidable if we accept the trend.
I think how things behave might depend to some extent on how you define an achieved time horizon; if there is a cost/speed requirement, then it becomes more plausible that longer horizon lengths would either have ~the same or lower reliability / success rate as smaller ones, once the AI surpasses humans in long-horizon agency. Similar to how if we created a version of HCAST but flipped based on AI times, then at a fixed speed budget human "reliability" might increase at higher time horizons, because our advantage is in long horizon agency and not speed.
In general things seem potentially sensitive to definitional choices and I don't feel like I've got things fully figured out in terms of what the behavior in the limit should be.
Thanks for writing this up! I actually mostly agree with everything you say about how much evidence the historical data points provide for a superexponential-given-no-automation trend. I think I place a bit more weight on it than you but I think we're close enough that it's not worth getting into.
The reason we have a superexponential option isn't primarily because of the existing empirical data, it's because we think the underlying curve is plausibly superexponential for conceptual reasons (in our original timelines forecast we had equal weight on superexponential and exponential, though after more thinking I'm considering giving more weight to superexponential). I think the current empirical evidence doesn't distinguish much between the two hypotheses.
In our latest published model we had an option for it being exponential up until a certain point, then becoming superexponential afterward. Though this seems fairly ad hoc so we might remove that in our next version.
The main thing I disagree with is your skepticism of the meta-skills argument, which is driving much of my credence. It just seems extremely unintuitive to me to think that it would take as much effective compute to go from 1 million years to 10 million years as it takes to go from 1 hour to 10 hours, so seems like we mainly have a difference in intuitions here. I agree it would be nice to make the argument more carefully, I won't take the time to try to do that right now. but will spew some more intuitions.
"We might expect progress in chess ability to be superexponential, as AIs start to figure out the meta-skills (such as tactical ability) required to fully understand how chess pieces can interact. That is, we would expect it to require more new skills to go from an ELO of 2400 to 2500, than it does to go from an ELO of 3400 to 3500."
I don't think this analogy as stated makes sense. My impression is that going from 3400 to 3500 is likely starting to bump up against the limits of how good you can be at Chess, or a weaker claims is that it is very superhuman. While we're talking about "just" reaching the level of a top human.
To my mind an AI that can do tasks that take top humans 1 million years feels like it's essentially top human level. And the same for 10 milion years, but very slightly better. So I'd think the equivalent of this jump is more like 2799.99 to 2799.991 ELO (2800 is roughly top human ELO). While the earlier 1 to 10 hour jump would be more like a 100 point jump or something.
I guess I'm just restating my intuitions that at higher levels the jump is a smaller of a difference in skills. I'm not sure how to further convey that. It personally feels like when I do a 1 hour vs. 10 hour programming task the latter often but not always involves significantly more high-level planning, investigating subtle bugs and consistently error correcting, etc.. While if I imagine spending 1 million years on a coding task, there's not really any new agency skills needed to get to 10 million years, I already have the ability to consistently make steady progress on a very difficult problem.
- My understanding is that AI 2027's forecast is heavily driven by putting substantial weight on such a superexponential fit, in which case my claim may call into question the reliability of this forecast. However, I have not dug into AI 2027's forecast, and am happy to be corrected on this point. My primary concern is with the specific claim I am making rather than how it relates to any particular aggregated forecast.
You can see a sensitivity analysis for our latest published model here, though we're working on a new one which might change things (the benchmarks and gaps model relies a lot less on the time horizon extrapolation which is why the difference is much smaller; also "superexponential immediately" is more aggressive than "becomes superexponential at some point" would be, due to the transition to superexponential that I mentioned above):
We're working on updating our timelines in a bunch of ways, and I've thought a bit more about this.
My current best guess, which isn't confident, is that if we took a version of the METR HCAST suite which didn't have any ambiguities or bugs, then the AGIs would have infinite time horizons. And that we should discuss this theoretical task suite rather than the literal HCAST in our timelines forecast, so therefore we should expect a finite time singularity. If we kept the ambiguities/bugs instead we'd have an asymptote and a non-infinite value, as would humans with inifinite time. In the paragraph you're quoting, I think that the main thing driving non-infinite values is that longer tasks are more likely to have ambiguities/bugs that make them unsolvable with full reliability.
I overall agree that things seem to be going slower than AI 2027 (and my median was longer when it came out).
However as mentioned in the caption, the green curve is a simplified version of our original timelines model. Apologies about that, I think it's reasonable to judge us based on that.
FWIW though, the central superexponential Mar 2027 trajectory from our original model certainly is not strongly contradicted by GPT-5, both with and without an AI R&D speedup interpolation issue fixed.
The original model, filtered for superexponential (pre-AI-R&D-automation) trajectories that reach superhuman coder in 2027:
With AI R&D speedup bug fixed, also filtered for superexponential pre-AI-R&D-automation (backcast looks much better, GPT-5 prediction slightly worse):
Either way, we're now working on a much improved model which will likely have an interactive web app which will provide an improvement over this static graph, e.g. you'll be able to try various parameter settings and see what time horizon trajectories they generate and how consistent they are with future data points.
Note also that the above trajectories are from the original model, not the May update model which we unfortunately aren't taking the time to create for various reasons, we think it would likely look a little worse in terms of the GPT-5 fit but might depend how you filter for which trajectories count as superexponential.
I (AI 2027 co-author) mostly agree (haven't run this by any other authors).
I agree regarding the GPT-5 release being something to pay attention to.
Whenever I put specific metrics in the text it was supposed to refer to the end of the time period (so "mid 2025" would be end of August since that's that's 2/3 through the year), but I never made that clear anywhere, my bad and I think the middle of the period is a natural interpretation as well. You can see that on the side panel on the right the date starts at Aug 2025.
But either way, waiting until the end of Aug probably won't change the directional update.
What these results do provide is evidence that, indeed, the suggested AI 2027 timelines are unlikely to be met; this should perhaps nudge us towards slightly longer timelines.
I agree that these results should nudge us toward slightly longer timelines, I think bolding that AI 2027 timelines are unlikely to be met seems like it's taking too much from these few early data points if you're implying that these play a big factor in that conclusion (though I might agreed with your assessment even before these data points depending on what you mean by unlikely).
As you allude to, the AI 2027 teams' current median timelines are all longer than 2027 to varying extents (roughly, 2028-2033). We're currently working on an update to our teams' timelines forecasts and will share an update on our views soon.
If the consensus is that the original timeline was too aggressive, another idea for community effort is rewriting the original plot with a longer-term horizon, say an "AI 2030" narrative, which will be published here and then analyzed in real-time, month by month.
We're in the early stage of working on something like this. It wouldn't be just the original plot then stretched out, it would be our best guess predictions conditional on longer timelines.
Okay, it sounds like our disagrement basically boils down to the value of the forecasts as well as the value of the scenario format (does that seem right?), which I don't think is something we'll come to agreement on.
Thanks again for writing this up! I hope you're right about timelines being much longer and 2027 being insane (as I mentioned, it's faster than my median has ever been, but I think it's plausible enough to take seriously).
edit: I'd also be curious for you to specify what you mean by academic? The scenario itself seems like a very unusual format for academia. I think it would have seemed more serious academic-y if we had ditched the scenario format.