Strong upvote, even though I think you're wrong about some important claims here, because you're being detailed enough for me to reply to.
... which I will do properly (ie, with citations into your post) tomorrow, if it still seems useful to be more specific. But the gist of what I'll defend better if needed is: I think while it's quite possible that the predictions function as OpenAI propaganda, that's separate from whether they're doing that because they are valuable - if someone had come up with these predictions in a box, isolated from OpenAI, it'd have similar effects; so then the question is separately about upstream causality of why say these things (credit and blame assignment), vs downstream causality of what these things will do (and what to do about it now). The upstream causality seems like a distraction, except inasmuch as it's relevant to downstream causality (eg, because properly assigned credit or blame might change the landscape of the present). IMO the main concern here is that these predictions, which were already being made by many people around the tech but not so specifically or with such careful argumentation, seem to be somehow being used by OpenAI to further their purposes. If that's because the predictions turn out correct, that maybe seems worse than if they were wrong, because they're pretty scary predictions - but either way, it's not good news that, in my view, there doesn't seem to be such a thing as bad publicity for AGI, and I still don't know for sure why that's happening. And that seems like where most of the value is in figuring out this discussion, to me, at least. Though the view you initially appeared to be writing down, that the predictions themselves are functioning as a propaganda piece in an upstream-causality intent sort of way. does seem to be a common one, so having a good and solid debate about it where we try to figure out and confirm the who-did-what-why a bit might well be worth the attention.
in my view, there doesn't seem to be such a thing as bad publicity for AGI, and I still don't know for sure why that's happening. And that seems like where most of the value is in figuring out this discussion, to me, at least.
It's an incentive problem.
There is no way to discuss something being dangerous that does not also render it valuable. People are incentivized to seek out value; our entire economy is based on it. It works beautifully, but it is terrible at mitigating externalities. We only dial back from dangerous or bad things after the disaster; so long as doing things is profitable, rational economic actors seek out high-risk activities as far as permitted, because they alone get the profit and the majority of the risk is to other people.
In my view Yudkowsky's body of work has had two main effects, which run in opposite directions:
Daniel is a thoughtful, strategic person who understands and thinks about AI strategy. He presumably wrote AI 2027 to try to influence strategy around AI. His perspective is going to be for playing as OpenAI. He will have used this perspective for years, totaling thousands of hours. He will have spent all of that time seeing AI research as a race, and trying to figure out how OpenAI can win. This is a generating function for OpenAI's investor pitch, and is also the perspective that AI 2027 takes.
S.K.'s comment: I would like to repeat the quote[1] from the AI-2027 forecast, which I first mentioned in another comment. "The scenario itself was written iteratively: we wrote the first period (up to mid-2025), then the following period, etc. until we reached the ending. We then scrapped this and did it again.
We weren’t trying to reach any particular ending. After we finished the first ending—which is now colored red—we wrote a new alternative branch because we wanted to also depict a more hopeful way things could end, starting from roughly the same premises. This went through several iterations".[2]
S.K.'s comment continues: In the unlikely event that it was DeepCent who aligned its AI and Consensus-1 ended up aligned, it would also be "a more hopeful way things could end". However, the story has OpenBrain AND DeepCent choose misaligning training environments and create misaligned AIs. The Slowdown Ending has OpenBrain retry solving alignment, this time with OOMs more effort. DeepCent, on the other hand, cannot retry without falling further behind.
Second: what information is available, and what information do you see a lot?
I think this is the main source of skew.
S.K.'s comment: the AI-2027 forecast relies on the following five pillars:
The forecast related to AI goals is unlikely to be skewed. Sections 1, 2 and 5 of the compute forecast don't actually rely[3] on the existence of powerful AIs. The security forecast is harder to grade, since it relies on humans deciding to guard the secrets against other humans, but also has benchmark-based estimates. What rests is the timelines forecast and the takeoff forecast.
The former one is so unreliable that even the authors acknowledged it in April 2025 by having Eli forecast the median date of superhuman coders' appearance to be 2027 (2025 to 2039), 2028 (2025 to >2050) or 2030 (2026 to >2050).
The takeoff forecast rests on the assumption that superhuman coders and AI researchers will appear and will greatly accelerate the AI research. The exact rates of acceleration are most vulnerable to being skewed, especially if the AIs are high-level neuralese before becoming superhuman coders. But I don't think that we even have better ways to forecast the acceleration.
For a concrete example of this that I didn't dig into in my review, from the AI 2027 timelines forecast.
We first show Method 1: time-horizon-extension, a relatively simple model which forecasts when SC will arrive by extending the trend established by METR’s report of AIs accomplishing tasks that take humans increasing amounts of time.
We then present Method 2: benchmarks-and-gaps, a more complex model starting from a forecast saturation of an AI R&D benchmark (RE-Bench), and then how long it will take to go from that system to one that can handle real-world tasks at the best AGI company.
Finally we then provide an “all-things-considered” forecast that takes into account these two models, as well as other possible influences such as geopolitics and macroeconomics.
Are either RE-Bench or the METR time horizon[4] metrics good metrics, as-is? Will they continue to extrapolate? Will a model that saturates them accelerate research a lot?
S.K.'s comment: the authors start "from a forecast saturation of an AI R&D benchmark (RE-Bench), and then [estimate] how long it will take to go from that system to one that can handle real-world tasks at the best AGI company". Saturating the RE-bench, unlike reaching the METR-like time horizon of a month, is, of course, NOT enough to accelerate AI research.
AI 2027’s “Vice President” (read: JD Vance) election subplot is long and also almost totally irrelevant to the plot. It is so conspicuously strange that I had trouble figuring out why it would even be there. I didn’t learn until after I’d written my take that JD Vance had read AI 2027 and mentioned it in an interview, which also seems like a very odd thing to happen. I went looking for the simplest explanation I could.
S.K.'s comment: the election-related subplot and mentions of Vance are due to the fact that 2028 is an election year in the USA. The American Constitution prohibits Trump from becoming the POTUS in 2028, so the Americans will have to choose between another Republican and a Democrat. The Republican candidate is most likely to be Vance.
Similarly, the line about Thiel getting the flying car could likely be a reference to a popular joke coined by Thiel: "We wanted flying cars, instead we got 140 characters."
S.K.'s footnote: The quote is found in the collapsible section "How did we write it?" on the forecast's main page.
S.K.'s footnote: the authors also claim that "It was overall more difficult, because unlike with the first ending, we were trying to get it to reach a good outcome starting from a rather difficult situation."
S.K.'s footnote: Sections 3 and 4 have powerful AIs used for automating research, but they require only 5% of OpenBrain's compute.
S.K.'s footnote: The METR benchmark has already run into issues with spurious failures and Grok 4's failure on fast tasks.
This is taken from a comment I wrote because it ended up being very long, and it addressed objections I have heard from multiple people. I include Neel's previous comment for context. Previous post here.[1]
Neel Nanda:
I haven't read the whole post, but the claims that this can be largely dismissed because of implicit bias towards the pro OpenAI narrative are completely ridiculous and ignorant of the background context of the authors. Most of the main authors of the piece have never worked at OpenAI or any other AGI lab. Daniel held broadly similar of use to this many years ago before he joined Openai. I know because he has both written about them and I had conversations with him before he joined openai where he expressed broadly similar views. I don't fully agree with these views, but they were detailed and well thought out and were a better prediction of the future than mine at the time. And he also was willing to sign away millions of dollars of equity in order to preserve his integrity - implying that him having OpenAI stock is causing him to warp his underlying beliefs seems an enormous stretch. And to my knowledge, AI 2027 did not receive any OpenPhil funding.
I find it frustrating and arrogant when people assume without good reason that disagreement is because of some background bias in the other person - often people disagree with you because of actual reasons!
These issues specifically have been a sticking point for a number of people, so I should clarify some things separately. Probably this is also because I didn't see this earlier so it's been a while and because I know who you are.
I do not think AI 2027 is, effectively, OpenAI's propaganda because it is about a recursively self-improving AI and OpenAI is also about RSI. There are a lot of versions (and possible versions) of a recursively self-improving AI thesis. Daniel Kokotajlo has been around long enough that he was definitely familiar with the territory before he worked at OpenAI. I think that it is effectively OpenAI propaganda because it assumes a very specific path to a recursively self-improving AI with a very specific technical, social and business environment, and this story is about a company that appears to closely resemble OpenAI[2] and is pursuing something very similar to OpenAI's current strategy. It seems unlikely that Daniel had these very specific views before he started at OpenAI in 2022.
Daniel is a thoughtful, strategic person who understands and thinks about AI strategy. He presumably wrote AI 2027 to try to influence strategy around AI. His perspective is going to be for playing as OpenAI. He will have used this perspective for years, totaling thousands of hours. He will have spent all of that time seeing AI research as a race, and trying to figure out how OpenAI can win. This is a generating function for OpenAI's investor pitch, and is also the perspective that AI 2027 takes.
Working at OpenAI means spending years of your professional life completely immersed in an information environment sponsored by, and meant to increase the value of, OpenAI. Having done that is a relevant factor for what information you think is true and what assumptions you think are reasonable. Even if you started off with few opinions about them, and you very critically examined and rejected most of what OpenAI said about itself internally, you would still have skewed perspective about OpenAI and things concerning OpenAI.
I think of industries I have worked in from the perspective of the company I worked for when I was in that industry. I expect that when he worked at OpenAI he was doing his best to figure out how OpenAI comes out ahead, and so was everyone around him. This would have been true whether or not he was being explicitly told to do it, and whether or not he was on the clock. It is simpler to expect that this did influence him than to expect that it did not.
Quitting OpenAI loudly doesn't really change this picture, because you generally only quit loudly if you have a specific bone to pick. If you've got a bone to pick while quitting OpenAI, that bone is, presumably, with OpenAI. Whatever story you tell after you do that is probably about OpenAI.
I think the part about financial incentives is getting dismissed sometimes because a lot of ill-informed people have tried to talk about finance in AI. This seems to have become sort of a thought-terminating cliche, where any question about the financial incentives around AI is assumed to be from uninformed people. I will try to explain what I meant about the financial influence in a little more detail.
In this specific case, I think that the authors are probably well-intentioned. However, most of their shaky assumptions just happen to be things which would be worth at least a hundred billion dollars to OpenAI specifically if they were true. If you were writing a pitch to try to get funding for OpenAI or a similar company, you would have billions of reasons to be as persuasive as possible about these things. Given the power of that financial incentive, it's not surprising that people have come up with compelling stories that just happen to make good investor pitches. Well-intentioned people can be so immersed in them that they cannot see past them.
It is worth noting that the lead author of AI 2027 is a former OpenAI employee. He is mostly famous outside OpenAI for having refused to sign their non-disparagement agreement and for advocating for stricter oversight of AI businesses. I do not think it is very credible that he is deliberately shilling for OpenAI here. I do think it is likely that he is completely unable to see outside their narrative, which they have an intense financial interest in sustaining.
There are a lot of different ways for a viewpoint to be skewed by money.
First is to just be paid to say things.
I don't think anyone was paid anything by OpenAI for writing AI 2027. I thought I made enough of a point of that in the article, but the second block above is towards the end of the relevant section and I should maybe have put it towards the top. I will remember to do that if I am writing something like this again and maybe make sure to write at least an extra paragraph or two about it.
I do not think Daniel is deliberately shilling for OpenAI. That's not an accusation I think is even remotely supportable, and in fact there's a lot of credible evidence running the other way. He's got a very long track record and he made a massive point of publicly dissenting from their non-disparagement agreement. It would take a lot of counter-evidence to convince me of his insincerity.
You didn't bring him up, but I also don't think Scott, who I think is responsible for most of the style of the piece, is being paid by anyone in particular to say anything in particular. I doubt such a thing is possible even in principle. Scott has a pretty solid track record of saying whatever he wants to say.
Second: what information is available, and what information do you see a lot?
I think this is the main source of skew.
If it's valuable to convince people something is true, you will probably look for facts and arguments which make it seem true. You will be less likely to look for facts and arguments which make it seem false. You will then make sure that as many people are aware of all the facts and arguments that make the thing seem true as possible.
At a corporate level this doesn't even have to be a specific person. People who are pursuing things that look promising for the company will be given time and space to pursue what they are doing, and people who are not will be more likely to be told to find something else to do. You will choose to promote favorable facts and not promote bad ones. You get the same effect as if a single person had deliberately chosen to only look for good facts.
It would be weird if this wasn't true of OpenAI given how much money is involved. As in, positively anomalous. You do not raise money by seeking out reasons why your technology is maybe not worth money, or by making sure everyone knows those things. Why would you do that? You are getting money, directly, because people think the technology you are working on is worth a lot of money, and everyone knows as much as you can give them about why what you're doing is worth a lot of money.
Tangentially, this type of narrative allows companies to convince staff to take compensation that is more heavily weighted towards stock, which tends to benefit existing shareholders in cases where they prefer to do that. They know employees will probably sell it back to them well below value at public sale or acquisition, or they know the stock is worth less than salary would be.
For a concrete example of this that I didn't dig into in my review, from the AI 2027 timelines forecast.
We first show Method 1: time-horizon-extension, a relatively simple model which forecasts when SC will arrive by extending the trend established by METR’s report of AIs accomplishing tasks that take humans increasing amounts of time.
We then present Method 2: benchmarks-and-gaps, a more complex model starting from a forecast saturation of an AI R&D benchmark (RE-Bench), and then how long it will take to go from that system to one that can handle real-world tasks at the best AGI company.
Finally we then provide an “all-things-considered” forecast that takes into account these two models, as well as other possible influences such as geopolitics and macroeconomics.
Are either RE-Bench or the METR time horizon metrics good metrics, as-is? Will they continue to extrapolate? Will a model that saturates them accelerate research a lot?
I think the answer to all of these is maybe. If you're OpenAI, it is pretty important that benchmarks are good metrics. It is worth a ton of money. So, institutionally, OpenAI has to believe in benchmarks, and vastly prefers if the answer is "yes" to all of these questions. And this is also what AI 2027 is assuming.
I made a point of running this point into the ground in writing it up, but essentially every time we break a "maybe" question in AI 2027, the answer seems to be the one that OpenAI is also likely to prefer. It's a very specific thing to happen! It doesn't seem very likely it happened by chance. In total the effect is that "this is a slight dissent from the OpenAI hype pitch", in my opinion.
This isn't even a problem entirely among OpenAI people. OpenAI has the loudest voice and is more or less setting the agenda for the industry. This is both because they were very clearly in the lead for a stretch, and because they've been very successful at acquiring users and raising money. There are probably more people who are bought into OpenAI's exact version of everything outside the company than inside of it. This is a considerable problem if you want a correct evaluation of the current trajectory.
I obviously cannot prove this, but I think if Daniel hadn't been a former OpenAI employee I probably would have basically the same criticism of the actual writing. It would be neater, even, because "this person has bought into OpenAI's hype" is a lot less complicated without the non-disparagement thing, which buys a lot of credibility. I honestly didn't want to mention who any of the authors were at all, but it seemed entirely too relevant to the case I was making to do it.
That's two: being paid and having skewed information.
Third thing, much smaller, just being slanted because you have a financial incentive. Maybe you’re just optimistic, maybe you’re hoping to sell soon.
Daniel probably still owns stock or options. I mentioned this in the piece. I don't think this is very relevant or is very likely to skew his perspective. It did seem like I would be failing to explain what was going on if I did not mention the possibility while discussing how he relates to OpenAI. I think it is incredibly weak evidence when stacked against his other history with the company, which strongly indicates that he's not inclined to lie for them or even be especially quiet when he disagrees with them.
I don't think it's disgraceful to mention that people have direct financial incentives. There's I think an implicit understanding that it's uncouth to mention this sort of thing, and I disagree with it. I think it causes severe problems, in general. People who own significant stock in companies shouldn't be assumed to be unbiased when discussing those companies, and it shouldn't be taboo to mention the potential slant.
My last point is stranger, and is only sort of about money. If everyone you know is financially involved, is there some point where you might as well be?
JD Vance gets flattered anonymously by describing him using his job title, but we flatter Peter Thiel by name. Peter Thiel is, actually, the only person who gets a shout-out by name. Maybe being an early investor in OpenAI is the only way to earn that. I didn’t previously suspect that he was the sole or primary donor funding the think tank that this came out of, but now I do. I am reminded that the second named author of this paper has a pretty funny post about how everyone doing something weird at all the parties he goes to is being bankrolled by Peter Thiel.
This is about Scott, mostly.
AI 2027’s “Vice President” (read: JD Vance) election subplot is long and also almost totally irrelevant to the plot. It is so conspicuously strange that I had trouble figuring out why it would even be there. I didn’t learn until after I’d written my take that JD Vance had read AI 2027 and mentioned it in an interview, which also seems like a very odd thing to happen. I went looking for the simplest explanation I could.
Scott says whatever he wants, but apparently by his accounting half of his social circle is being bankrolled by Peter Thiel. This part of AI 2027 seems to be him, and he seems to be deliberately flattering Vance. Vance is a pretty well known Thiel acolyte. On the relatively happy ending of AI 2027 they build an ASI surveillance system, and surveillance is a big Peter Thiel hobby horse.
I don't know what I'm really supposed to make of any of this. I definitely noticed it. It raises a lot of questions. It definitely seems to suggest strongly that if you spend a decade or more bankrolling all of Scott's friends to do weird things they think are interesting, you are likely to see Scott flatter you and your opinions in writing. It also seems to suggest that Scott's deliberately acting to lobby JD Vance. If it weren't for Peter Thiel bankrolling his friends so much that Scott makes a joke out of it, I would think it just looked like Scott had a very Thiel-adjacent friend group.
In pretty much the same way that OpenAI will tend to generate pro-OpenAI facts and arguments, and not generate anti-OpenAI facts and arguments, I would expect that if enough people around you are being bankrolled by someone for long enough they will tend to produce information that person likes and not produce information that person doesn't like.
I cannot find a simpler explanation than Thiel influence for why you would have a reasonably long subplot about JD Vance, world domination, and mass surveillance and then mention Peter Thiel in the finale.
I don't think pointing out this specific type of connection should be taboo for basically the same reason I don't think pointing out who owns what stock should be. I like knowing things, and being correct about them, and so I like knowing if people are offering good information or if there is an obvious reason their information or arguments would be bad.
If making a proper post out of a very long comment like this is considered poor form, I claim ignorance.
A few people have said that it could be DeepMind. I think it could be but pretty clearly isn't. Among other things, DeepMind would not want or need to sell products they considered dangerous or to be possibly close to allowing RSI, because they are extremely cash-rich. If the forecast were about DeepMind, it would probably consider this, but it isn't, so it doesn't.